var/home/core/zuul-output/0000755000175000017500000000000015113332336014525 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015113345156015475 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005105210115113345147017674 0ustar rootrootDec 01 15:34:57 crc systemd[1]: Starting Kubernetes Kubelet... Dec 01 15:34:57 crc restorecon[4738]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:57 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 15:34:58 crc restorecon[4738]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 15:34:58 crc restorecon[4738]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 01 15:34:58 crc kubenswrapper[4739]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 15:34:58 crc kubenswrapper[4739]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 01 15:34:58 crc kubenswrapper[4739]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 15:34:58 crc kubenswrapper[4739]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 15:34:58 crc kubenswrapper[4739]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 01 15:34:58 crc kubenswrapper[4739]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.278614 4739 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285340 4739 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285644 4739 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285654 4739 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285662 4739 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285670 4739 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285679 4739 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285687 4739 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285696 4739 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285704 4739 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285711 4739 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285719 4739 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285727 4739 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285735 4739 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285744 4739 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285769 4739 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285777 4739 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285785 4739 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285794 4739 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285802 4739 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285809 4739 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285818 4739 feature_gate.go:330] unrecognized feature gate: Example Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285842 4739 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285851 4739 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285858 4739 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285866 4739 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285873 4739 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285885 4739 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285895 4739 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285906 4739 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285915 4739 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285923 4739 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285932 4739 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285942 4739 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285951 4739 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285972 4739 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285981 4739 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.285990 4739 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286000 4739 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286009 4739 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286018 4739 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286026 4739 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286036 4739 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286045 4739 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286053 4739 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286062 4739 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286072 4739 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286082 4739 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286091 4739 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286098 4739 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286106 4739 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286113 4739 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286121 4739 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286129 4739 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286136 4739 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286144 4739 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286153 4739 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286163 4739 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286172 4739 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286180 4739 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286187 4739 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286195 4739 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286203 4739 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286210 4739 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286218 4739 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286225 4739 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286233 4739 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286240 4739 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286248 4739 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286256 4739 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286264 4739 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.286271 4739 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286722 4739 flags.go:64] FLAG: --address="0.0.0.0" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286746 4739 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286783 4739 flags.go:64] FLAG: --anonymous-auth="true" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286794 4739 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286805 4739 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286814 4739 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286826 4739 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286837 4739 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286847 4739 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286856 4739 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286866 4739 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286887 4739 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286898 4739 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286907 4739 flags.go:64] FLAG: --cgroup-root="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286916 4739 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286924 4739 flags.go:64] FLAG: --client-ca-file="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286934 4739 flags.go:64] FLAG: --cloud-config="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286943 4739 flags.go:64] FLAG: --cloud-provider="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286952 4739 flags.go:64] FLAG: --cluster-dns="[]" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286969 4739 flags.go:64] FLAG: --cluster-domain="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286978 4739 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286987 4739 flags.go:64] FLAG: --config-dir="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.286996 4739 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287006 4739 flags.go:64] FLAG: --container-log-max-files="5" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287017 4739 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287027 4739 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287036 4739 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287045 4739 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287055 4739 flags.go:64] FLAG: --contention-profiling="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287064 4739 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287073 4739 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287082 4739 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287091 4739 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287102 4739 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287111 4739 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287120 4739 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287129 4739 flags.go:64] FLAG: --enable-load-reader="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287138 4739 flags.go:64] FLAG: --enable-server="true" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287147 4739 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287165 4739 flags.go:64] FLAG: --event-burst="100" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287174 4739 flags.go:64] FLAG: --event-qps="50" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287183 4739 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287193 4739 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287202 4739 flags.go:64] FLAG: --eviction-hard="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287213 4739 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287222 4739 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287231 4739 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287252 4739 flags.go:64] FLAG: --eviction-soft="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287261 4739 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287270 4739 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287280 4739 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287289 4739 flags.go:64] FLAG: --experimental-mounter-path="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287297 4739 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287306 4739 flags.go:64] FLAG: --fail-swap-on="true" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287316 4739 flags.go:64] FLAG: --feature-gates="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287326 4739 flags.go:64] FLAG: --file-check-frequency="20s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287335 4739 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287344 4739 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287353 4739 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287363 4739 flags.go:64] FLAG: --healthz-port="10248" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287372 4739 flags.go:64] FLAG: --help="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287381 4739 flags.go:64] FLAG: --hostname-override="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287390 4739 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287399 4739 flags.go:64] FLAG: --http-check-frequency="20s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287408 4739 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287448 4739 flags.go:64] FLAG: --image-credential-provider-config="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287457 4739 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287466 4739 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287475 4739 flags.go:64] FLAG: --image-service-endpoint="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287484 4739 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287493 4739 flags.go:64] FLAG: --kube-api-burst="100" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287502 4739 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287514 4739 flags.go:64] FLAG: --kube-api-qps="50" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287522 4739 flags.go:64] FLAG: --kube-reserved="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287532 4739 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287541 4739 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287550 4739 flags.go:64] FLAG: --kubelet-cgroups="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287559 4739 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287568 4739 flags.go:64] FLAG: --lock-file="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287576 4739 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287586 4739 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287596 4739 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287609 4739 flags.go:64] FLAG: --log-json-split-stream="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287630 4739 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287640 4739 flags.go:64] FLAG: --log-text-split-stream="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287649 4739 flags.go:64] FLAG: --logging-format="text" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287658 4739 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287668 4739 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287677 4739 flags.go:64] FLAG: --manifest-url="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287686 4739 flags.go:64] FLAG: --manifest-url-header="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287697 4739 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287706 4739 flags.go:64] FLAG: --max-open-files="1000000" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287717 4739 flags.go:64] FLAG: --max-pods="110" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287726 4739 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287735 4739 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287744 4739 flags.go:64] FLAG: --memory-manager-policy="None" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287753 4739 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287762 4739 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287771 4739 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287781 4739 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287801 4739 flags.go:64] FLAG: --node-status-max-images="50" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287810 4739 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287819 4739 flags.go:64] FLAG: --oom-score-adj="-999" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287828 4739 flags.go:64] FLAG: --pod-cidr="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287837 4739 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287850 4739 flags.go:64] FLAG: --pod-manifest-path="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287859 4739 flags.go:64] FLAG: --pod-max-pids="-1" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287868 4739 flags.go:64] FLAG: --pods-per-core="0" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287877 4739 flags.go:64] FLAG: --port="10250" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287887 4739 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287895 4739 flags.go:64] FLAG: --provider-id="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287904 4739 flags.go:64] FLAG: --qos-reserved="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287913 4739 flags.go:64] FLAG: --read-only-port="10255" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287923 4739 flags.go:64] FLAG: --register-node="true" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287932 4739 flags.go:64] FLAG: --register-schedulable="true" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287941 4739 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287955 4739 flags.go:64] FLAG: --registry-burst="10" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287964 4739 flags.go:64] FLAG: --registry-qps="5" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287973 4739 flags.go:64] FLAG: --reserved-cpus="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.287994 4739 flags.go:64] FLAG: --reserved-memory="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288005 4739 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288015 4739 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288024 4739 flags.go:64] FLAG: --rotate-certificates="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288033 4739 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288042 4739 flags.go:64] FLAG: --runonce="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288051 4739 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288060 4739 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288069 4739 flags.go:64] FLAG: --seccomp-default="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288078 4739 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288087 4739 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288097 4739 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288106 4739 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288116 4739 flags.go:64] FLAG: --storage-driver-password="root" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288124 4739 flags.go:64] FLAG: --storage-driver-secure="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288133 4739 flags.go:64] FLAG: --storage-driver-table="stats" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288142 4739 flags.go:64] FLAG: --storage-driver-user="root" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288151 4739 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288161 4739 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288170 4739 flags.go:64] FLAG: --system-cgroups="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288178 4739 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288192 4739 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288201 4739 flags.go:64] FLAG: --tls-cert-file="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288210 4739 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288226 4739 flags.go:64] FLAG: --tls-min-version="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288235 4739 flags.go:64] FLAG: --tls-private-key-file="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288249 4739 flags.go:64] FLAG: --topology-manager-policy="none" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288257 4739 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288267 4739 flags.go:64] FLAG: --topology-manager-scope="container" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288276 4739 flags.go:64] FLAG: --v="2" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288287 4739 flags.go:64] FLAG: --version="false" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288298 4739 flags.go:64] FLAG: --vmodule="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288308 4739 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.288318 4739 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288553 4739 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288565 4739 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288601 4739 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288610 4739 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288619 4739 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288628 4739 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288637 4739 feature_gate.go:330] unrecognized feature gate: Example Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288645 4739 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288654 4739 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288662 4739 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288670 4739 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288679 4739 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288687 4739 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288696 4739 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288704 4739 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288712 4739 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288720 4739 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288728 4739 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288735 4739 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288743 4739 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288750 4739 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288758 4739 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288766 4739 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288774 4739 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288785 4739 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288793 4739 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288803 4739 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288812 4739 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288821 4739 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288829 4739 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288836 4739 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288844 4739 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288852 4739 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288860 4739 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288868 4739 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288875 4739 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288884 4739 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288892 4739 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288911 4739 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288919 4739 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288927 4739 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288935 4739 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288942 4739 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288953 4739 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288961 4739 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288969 4739 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288976 4739 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288987 4739 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.288996 4739 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289005 4739 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289014 4739 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289022 4739 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289030 4739 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289038 4739 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289048 4739 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289058 4739 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289070 4739 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289078 4739 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289087 4739 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289095 4739 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289102 4739 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289110 4739 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289118 4739 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289126 4739 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289133 4739 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289141 4739 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289151 4739 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289159 4739 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289167 4739 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289175 4739 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.289182 4739 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.289455 4739 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.301168 4739 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.301208 4739 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301360 4739 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301382 4739 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301392 4739 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301401 4739 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301413 4739 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301449 4739 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301457 4739 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301466 4739 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301474 4739 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301485 4739 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301496 4739 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301506 4739 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301514 4739 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301523 4739 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301531 4739 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301539 4739 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301548 4739 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301558 4739 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301570 4739 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301581 4739 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301590 4739 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301600 4739 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301609 4739 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301620 4739 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301628 4739 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301637 4739 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301645 4739 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301653 4739 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301661 4739 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301671 4739 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301679 4739 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301689 4739 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301698 4739 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301706 4739 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301714 4739 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301721 4739 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301729 4739 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301737 4739 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301745 4739 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301753 4739 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301761 4739 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301768 4739 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301776 4739 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301783 4739 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301792 4739 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301799 4739 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301807 4739 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301815 4739 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301822 4739 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301859 4739 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301867 4739 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301875 4739 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301883 4739 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301891 4739 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301899 4739 feature_gate.go:330] unrecognized feature gate: Example Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301908 4739 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301916 4739 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301924 4739 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301931 4739 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301939 4739 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301947 4739 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301954 4739 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301962 4739 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301969 4739 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301977 4739 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301986 4739 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.301994 4739 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302002 4739 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302009 4739 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302017 4739 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302025 4739 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.302038 4739 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302256 4739 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302268 4739 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302277 4739 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302286 4739 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302293 4739 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302301 4739 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302309 4739 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302317 4739 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302325 4739 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302332 4739 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302340 4739 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302348 4739 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302355 4739 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302365 4739 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302376 4739 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302387 4739 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302396 4739 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302405 4739 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302413 4739 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302442 4739 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302450 4739 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302458 4739 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302466 4739 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302473 4739 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302482 4739 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302489 4739 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302497 4739 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302504 4739 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302512 4739 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302521 4739 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302529 4739 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302537 4739 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302545 4739 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302553 4739 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302561 4739 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302569 4739 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302580 4739 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302590 4739 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302598 4739 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302607 4739 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302616 4739 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302625 4739 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302633 4739 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302643 4739 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302652 4739 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302661 4739 feature_gate.go:330] unrecognized feature gate: Example Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302669 4739 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302677 4739 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302687 4739 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302695 4739 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302703 4739 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302711 4739 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302721 4739 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302729 4739 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302737 4739 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302745 4739 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302753 4739 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302760 4739 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302768 4739 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302775 4739 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302783 4739 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302791 4739 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302799 4739 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302807 4739 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302815 4739 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302823 4739 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302832 4739 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302839 4739 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302847 4739 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302854 4739 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.302862 4739 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.302874 4739 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.303064 4739 server.go:940] "Client rotation is on, will bootstrap in background" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.307047 4739 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.307194 4739 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.308030 4739 server.go:997] "Starting client certificate rotation" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.308058 4739 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.308640 4739 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-29 00:53:35.660846133 +0000 UTC Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.308736 4739 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 657h18m37.352114374s for next certificate rotation Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.315345 4739 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.318122 4739 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.333474 4739 log.go:25] "Validated CRI v1 runtime API" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.355382 4739 log.go:25] "Validated CRI v1 image API" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.357408 4739 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.361089 4739 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-01-15-30-44-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.361137 4739 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.393461 4739 manager.go:217] Machine: {Timestamp:2025-12-01 15:34:58.391336122 +0000 UTC m=+0.217082296 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:35294ef1-eb1e-44ab-9a6d-9a0c5248b388 BootID:d50300d8-e6ab-4451-94a5-1b5cdce96a01 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:3b:6e:7c Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:3b:6e:7c Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:2c:72:2a Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:f8:b1:bf Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:37:c2:81 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:b5:33:d3 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:5a:6c:61 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:4a:04:29:ba:5b:3a Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:82:58:4d:c5:98:0b Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.393836 4739 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.394039 4739 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.394442 4739 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.394798 4739 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.394856 4739 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.395311 4739 topology_manager.go:138] "Creating topology manager with none policy" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.395331 4739 container_manager_linux.go:303] "Creating device plugin manager" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.395595 4739 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.395658 4739 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.396059 4739 state_mem.go:36] "Initialized new in-memory state store" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.396184 4739 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.397106 4739 kubelet.go:418] "Attempting to sync node with API server" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.397136 4739 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.397159 4739 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.397180 4739 kubelet.go:324] "Adding apiserver pod source" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.397198 4739 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.399552 4739 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.399987 4739 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.400824 4739 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.400997 4739 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.106:6443: connect: connection refused Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.401356 4739 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.106:6443: connect: connection refused Dec 01 15:34:58 crc kubenswrapper[4739]: E1201 15:34:58.401530 4739 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.106:6443: connect: connection refused" logger="UnhandledError" Dec 01 15:34:58 crc kubenswrapper[4739]: E1201 15:34:58.401560 4739 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.106:6443: connect: connection refused" logger="UnhandledError" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.401837 4739 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.401907 4739 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.401920 4739 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.401932 4739 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.401950 4739 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.401973 4739 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.401986 4739 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.402002 4739 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.402015 4739 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.402027 4739 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.402045 4739 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.402058 4739 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.402369 4739 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.403849 4739 server.go:1280] "Started kubelet" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.403852 4739 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.106:6443: connect: connection refused Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.404515 4739 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.404653 4739 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.405642 4739 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 01 15:34:58 crc systemd[1]: Started Kubernetes Kubelet. Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.409402 4739 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.409499 4739 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.409590 4739 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.409606 4739 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 01 15:34:58 crc kubenswrapper[4739]: E1201 15:34:58.409700 4739 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.409557 4739 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 10:47:42.862770344 +0000 UTC Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.409783 4739 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1075h12m44.45299778s for next certificate rotation Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.409741 4739 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 01 15:34:58 crc kubenswrapper[4739]: E1201 15:34:58.410504 4739 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" interval="200ms" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.410569 4739 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.410589 4739 factory.go:55] Registering systemd factory Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.410601 4739 factory.go:221] Registration of the systemd container factory successfully Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.411252 4739 factory.go:153] Registering CRI-O factory Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.411279 4739 factory.go:221] Registration of the crio container factory successfully Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.411306 4739 factory.go:103] Registering Raw factory Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.411326 4739 manager.go:1196] Started watching for new ooms in manager Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.411325 4739 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.106:6443: connect: connection refused Dec 01 15:34:58 crc kubenswrapper[4739]: E1201 15:34:58.411412 4739 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.106:6443: connect: connection refused" logger="UnhandledError" Dec 01 15:34:58 crc kubenswrapper[4739]: E1201 15:34:58.411502 4739 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.106:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187d2159c2d74637 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 15:34:58.403796535 +0000 UTC m=+0.229542669,LastTimestamp:2025-12-01 15:34:58.403796535 +0000 UTC m=+0.229542669,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.412032 4739 manager.go:319] Starting recovery of all containers Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.415810 4739 server.go:460] "Adding debug handlers to kubelet server" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421459 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421531 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421546 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421568 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421585 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421608 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421626 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421644 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421674 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421692 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421719 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421738 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421760 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421786 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421813 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421831 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421926 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421948 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421961 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421981 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.421994 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422008 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422026 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422041 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422059 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422074 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422098 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422115 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422134 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422148 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422167 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422180 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422199 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422213 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422227 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422245 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422270 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422287 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422311 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422331 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422354 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422374 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422394 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422442 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422465 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422490 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422511 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422529 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422552 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422570 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422591 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422606 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422638 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422658 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422684 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422712 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422734 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.422756 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423326 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423382 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423443 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423468 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423489 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423518 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423542 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423571 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423591 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423613 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423644 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423665 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423694 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423715 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423737 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423765 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423788 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423815 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423836 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423858 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423888 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423910 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423942 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423967 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.423993 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424023 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424051 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424134 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424312 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424335 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424363 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424387 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424440 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424463 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424485 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424514 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424537 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424567 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424590 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424610 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424637 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424660 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424688 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424709 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424731 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424758 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424801 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424829 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424863 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424896 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424919 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424950 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.424976 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.425008 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.425038 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.425061 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.425092 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.425112 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.425301 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.425331 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.425354 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.425489 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.427755 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.427992 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.428105 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.428144 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.428241 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.430643 4739 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.430705 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.430730 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.430751 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.430774 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.430793 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.430811 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.430830 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.430849 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.430867 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.430935 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.430954 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.430972 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.430993 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431051 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431070 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431088 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431104 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431122 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431141 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431158 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431177 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431195 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431213 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431231 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431248 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431266 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431297 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431318 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431336 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431356 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431375 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431395 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431414 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431535 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431559 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431580 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431600 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431618 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431638 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431657 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431675 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431694 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431717 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431735 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431754 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431771 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431798 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431818 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431837 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431857 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431877 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431896 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431914 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431977 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.431996 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432019 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432042 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432062 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432084 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432105 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432154 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432172 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432190 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432209 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432227 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432245 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432264 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432282 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432299 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432318 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432334 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432351 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432370 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432387 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432406 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432452 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432471 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432489 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432506 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432526 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432543 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432560 4739 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432580 4739 reconstruct.go:97] "Volume reconstruction finished" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.432593 4739 reconciler.go:26] "Reconciler: start to sync state" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.436727 4739 manager.go:324] Recovery completed Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.448753 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.451763 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.451828 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.451844 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.453059 4739 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.453080 4739 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.453132 4739 state_mem.go:36] "Initialized new in-memory state store" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.466977 4739 policy_none.go:49] "None policy: Start" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.468391 4739 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.468475 4739 state_mem.go:35] "Initializing new in-memory state store" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.472651 4739 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.475805 4739 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.475844 4739 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.475904 4739 kubelet.go:2335] "Starting kubelet main sync loop" Dec 01 15:34:58 crc kubenswrapper[4739]: E1201 15:34:58.476053 4739 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.477087 4739 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.106:6443: connect: connection refused Dec 01 15:34:58 crc kubenswrapper[4739]: E1201 15:34:58.477158 4739 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.106:6443: connect: connection refused" logger="UnhandledError" Dec 01 15:34:58 crc kubenswrapper[4739]: E1201 15:34:58.509830 4739 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.538803 4739 manager.go:334] "Starting Device Plugin manager" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.538875 4739 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.538891 4739 server.go:79] "Starting device plugin registration server" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.539558 4739 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.539603 4739 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.540037 4739 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.540184 4739 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.540202 4739 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 01 15:34:58 crc kubenswrapper[4739]: E1201 15:34:58.549840 4739 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.576186 4739 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.576293 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.577533 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.577577 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.577589 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.577736 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.578030 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.578077 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.578611 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.578644 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.578656 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.578750 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.578899 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.578967 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.579396 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.579471 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.579487 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.579509 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.579495 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.579596 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.579745 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.579846 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.579892 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.581168 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.581202 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.581209 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.581228 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.581239 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.581215 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.581302 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.581335 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.581348 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.582095 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.582137 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.582206 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.582977 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.583006 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.583029 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.583293 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.583351 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.585257 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.585286 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.585298 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.585394 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.585440 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.585456 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:58 crc kubenswrapper[4739]: E1201 15:34:58.611479 4739 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" interval="400ms" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.634690 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.634822 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.634923 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.634994 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.635065 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.635112 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.635230 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.635281 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.635330 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.635455 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.635510 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.635565 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.635612 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.635662 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.635704 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.640268 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.641733 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.641775 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.641790 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.641816 4739 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 15:34:58 crc kubenswrapper[4739]: E1201 15:34:58.642343 4739 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.106:6443: connect: connection refused" node="crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.736692 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.736768 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.736817 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.736856 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.736896 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.736935 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.736980 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.736959 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.736939 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737039 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737089 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737091 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737153 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737175 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737130 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737225 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737252 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737271 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737322 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737334 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737348 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737373 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737382 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737468 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737481 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737517 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737540 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737559 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737612 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.737767 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.842571 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.844389 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.844480 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.844501 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.844536 4739 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 15:34:58 crc kubenswrapper[4739]: E1201 15:34:58.845077 4739 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.106:6443: connect: connection refused" node="crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.924122 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.931518 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.955862 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.957237 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-4376a04a3291be6b1ee889f58cd3d8214e01dff049aa96d290f867fcfb648265 WatchSource:0}: Error finding container 4376a04a3291be6b1ee889f58cd3d8214e01dff049aa96d290f867fcfb648265: Status 404 returned error can't find the container with id 4376a04a3291be6b1ee889f58cd3d8214e01dff049aa96d290f867fcfb648265 Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.973725 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:34:58 crc kubenswrapper[4739]: W1201 15:34:58.974352 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-01775a4492dc5a828a9fdde56aca8ff9c3c0d40ffad3bb25b403cae1356237d4 WatchSource:0}: Error finding container 01775a4492dc5a828a9fdde56aca8ff9c3c0d40ffad3bb25b403cae1356237d4: Status 404 returned error can't find the container with id 01775a4492dc5a828a9fdde56aca8ff9c3c0d40ffad3bb25b403cae1356237d4 Dec 01 15:34:58 crc kubenswrapper[4739]: I1201 15:34:58.982342 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 15:34:59 crc kubenswrapper[4739]: W1201 15:34:59.005703 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-de50122ce61286f3b56eb0668c39649b020ea8443444a5fc35b3aee40c1043a7 WatchSource:0}: Error finding container de50122ce61286f3b56eb0668c39649b020ea8443444a5fc35b3aee40c1043a7: Status 404 returned error can't find the container with id de50122ce61286f3b56eb0668c39649b020ea8443444a5fc35b3aee40c1043a7 Dec 01 15:34:59 crc kubenswrapper[4739]: E1201 15:34:59.012761 4739 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" interval="800ms" Dec 01 15:34:59 crc kubenswrapper[4739]: W1201 15:34:59.018293 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-4a8f2bc33bdcc0d0c72dcf5b061f1cc3fc4e76f8cc8fbdf5c35cbe13c182bc44 WatchSource:0}: Error finding container 4a8f2bc33bdcc0d0c72dcf5b061f1cc3fc4e76f8cc8fbdf5c35cbe13c182bc44: Status 404 returned error can't find the container with id 4a8f2bc33bdcc0d0c72dcf5b061f1cc3fc4e76f8cc8fbdf5c35cbe13c182bc44 Dec 01 15:34:59 crc kubenswrapper[4739]: W1201 15:34:59.244914 4739 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.106:6443: connect: connection refused Dec 01 15:34:59 crc kubenswrapper[4739]: E1201 15:34:59.245058 4739 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.106:6443: connect: connection refused" logger="UnhandledError" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.245721 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.247737 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.247788 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.247803 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.247847 4739 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 15:34:59 crc kubenswrapper[4739]: E1201 15:34:59.248550 4739 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.106:6443: connect: connection refused" node="crc" Dec 01 15:34:59 crc kubenswrapper[4739]: W1201 15:34:59.308621 4739 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.106:6443: connect: connection refused Dec 01 15:34:59 crc kubenswrapper[4739]: E1201 15:34:59.308749 4739 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.106:6443: connect: connection refused" logger="UnhandledError" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.405539 4739 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.106:6443: connect: connection refused Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.483252 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a"} Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.483797 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"de50122ce61286f3b56eb0668c39649b020ea8443444a5fc35b3aee40c1043a7"} Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.484995 4739 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589" exitCode=0 Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.485101 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589"} Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.485163 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"01775a4492dc5a828a9fdde56aca8ff9c3c0d40ffad3bb25b403cae1356237d4"} Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.485307 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.486762 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.486798 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.486811 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.486891 4739 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089" exitCode=0 Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.487006 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089"} Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.487071 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0710408a0300badcfe0ed835bcd9565b1f32c245668541868ccdecf10c255c57"} Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.487185 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.487994 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.488026 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.488041 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.488126 4739 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="9b31738b9fc87d330a787b510347c3e96279a085114fccc6d231b87233e7eced" exitCode=0 Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.488183 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"9b31738b9fc87d330a787b510347c3e96279a085114fccc6d231b87233e7eced"} Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.488216 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4376a04a3291be6b1ee889f58cd3d8214e01dff049aa96d290f867fcfb648265"} Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.488286 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.489305 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.489333 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.489345 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.489874 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.489908 4739 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477" exitCode=0 Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.489938 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477"} Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.489954 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4a8f2bc33bdcc0d0c72dcf5b061f1cc3fc4e76f8cc8fbdf5c35cbe13c182bc44"} Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.490017 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.490886 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.490915 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.490930 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.492082 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.492114 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:34:59 crc kubenswrapper[4739]: I1201 15:34:59.492127 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:34:59 crc kubenswrapper[4739]: W1201 15:34:59.513916 4739 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.106:6443: connect: connection refused Dec 01 15:34:59 crc kubenswrapper[4739]: E1201 15:34:59.514016 4739 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.106:6443: connect: connection refused" logger="UnhandledError" Dec 01 15:34:59 crc kubenswrapper[4739]: E1201 15:34:59.813628 4739 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" interval="1.6s" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.049217 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.051045 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.051092 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.051103 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.051135 4739 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 15:35:00 crc kubenswrapper[4739]: E1201 15:35:00.051624 4739 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.106:6443: connect: connection refused" node="crc" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.498879 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4"} Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.498972 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe"} Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.499018 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786"} Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.499043 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7"} Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.499068 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da"} Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.498980 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.500740 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.500780 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.500792 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.501776 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"741ce01ef8ea8b4716ee843f9642792ca701fa4e7af0b858817e43f02aa8c6d5"} Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.501931 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.503384 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.503448 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.503467 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.507639 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"da43d8d79a40a66229d0114f6276814c2cdcfbe049eb2a57e89fbb4076154c26"} Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.507705 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7d52f33a475d6eb938823fd4b7eac6ca8c026b3826c15bb8b67a50e7a6d942be"} Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.507729 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0adb30e972ba4be0f2150d59de63fc2042ebce799ebc7fe64faa9876ce0009b6"} Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.507859 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.508785 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.508807 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.508816 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.510271 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.510292 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d"} Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.510366 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081"} Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.510388 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13"} Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.512192 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.512218 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.512227 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.513447 4739 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0" exitCode=0 Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.513501 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0"} Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.513666 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.515154 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.515190 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.515202 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:00 crc kubenswrapper[4739]: I1201 15:35:00.962642 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.262098 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.325394 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.520165 4739 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631" exitCode=0 Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.520325 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.520338 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.520377 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.520493 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631"} Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.520672 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.521884 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.521925 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.521922 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.521973 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.521999 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.521943 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.522080 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.522109 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.522130 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.523618 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.523686 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.523707 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.652200 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.654000 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.654055 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.654074 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:01 crc kubenswrapper[4739]: I1201 15:35:01.654111 4739 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 15:35:02 crc kubenswrapper[4739]: I1201 15:35:02.528505 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb"} Dec 01 15:35:02 crc kubenswrapper[4739]: I1201 15:35:02.528589 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c"} Dec 01 15:35:02 crc kubenswrapper[4739]: I1201 15:35:02.528615 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e"} Dec 01 15:35:02 crc kubenswrapper[4739]: I1201 15:35:02.528631 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:02 crc kubenswrapper[4739]: I1201 15:35:02.528750 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:02 crc kubenswrapper[4739]: I1201 15:35:02.529890 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:02 crc kubenswrapper[4739]: I1201 15:35:02.529942 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:02 crc kubenswrapper[4739]: I1201 15:35:02.529961 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:02 crc kubenswrapper[4739]: I1201 15:35:02.530201 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:02 crc kubenswrapper[4739]: I1201 15:35:02.530268 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:02 crc kubenswrapper[4739]: I1201 15:35:02.530294 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:02 crc kubenswrapper[4739]: I1201 15:35:02.594868 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:35:02 crc kubenswrapper[4739]: I1201 15:35:02.834388 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:35:02 crc kubenswrapper[4739]: I1201 15:35:02.897586 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:35:03 crc kubenswrapper[4739]: I1201 15:35:03.536819 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b"} Dec 01 15:35:03 crc kubenswrapper[4739]: I1201 15:35:03.536886 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63"} Dec 01 15:35:03 crc kubenswrapper[4739]: I1201 15:35:03.536907 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:03 crc kubenswrapper[4739]: I1201 15:35:03.536936 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:03 crc kubenswrapper[4739]: I1201 15:35:03.537052 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:03 crc kubenswrapper[4739]: I1201 15:35:03.538129 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:03 crc kubenswrapper[4739]: I1201 15:35:03.538186 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:03 crc kubenswrapper[4739]: I1201 15:35:03.538200 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:03 crc kubenswrapper[4739]: I1201 15:35:03.538769 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:03 crc kubenswrapper[4739]: I1201 15:35:03.538805 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:03 crc kubenswrapper[4739]: I1201 15:35:03.538861 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:03 crc kubenswrapper[4739]: I1201 15:35:03.538879 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:03 crc kubenswrapper[4739]: I1201 15:35:03.538823 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:03 crc kubenswrapper[4739]: I1201 15:35:03.539082 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:04 crc kubenswrapper[4739]: I1201 15:35:04.539806 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:04 crc kubenswrapper[4739]: I1201 15:35:04.539859 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:04 crc kubenswrapper[4739]: I1201 15:35:04.541467 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:04 crc kubenswrapper[4739]: I1201 15:35:04.541523 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:04 crc kubenswrapper[4739]: I1201 15:35:04.541544 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:04 crc kubenswrapper[4739]: I1201 15:35:04.541669 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:04 crc kubenswrapper[4739]: I1201 15:35:04.541708 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:04 crc kubenswrapper[4739]: I1201 15:35:04.541725 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:06 crc kubenswrapper[4739]: I1201 15:35:06.145411 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:35:06 crc kubenswrapper[4739]: I1201 15:35:06.146184 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:06 crc kubenswrapper[4739]: I1201 15:35:06.147916 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:06 crc kubenswrapper[4739]: I1201 15:35:06.147963 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:06 crc kubenswrapper[4739]: I1201 15:35:06.147979 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:06 crc kubenswrapper[4739]: I1201 15:35:06.153504 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:35:06 crc kubenswrapper[4739]: I1201 15:35:06.546289 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:06 crc kubenswrapper[4739]: I1201 15:35:06.547835 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:06 crc kubenswrapper[4739]: I1201 15:35:06.547896 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:06 crc kubenswrapper[4739]: I1201 15:35:06.547915 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:07 crc kubenswrapper[4739]: I1201 15:35:07.386285 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:35:07 crc kubenswrapper[4739]: I1201 15:35:07.548774 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:07 crc kubenswrapper[4739]: I1201 15:35:07.550820 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:07 crc kubenswrapper[4739]: I1201 15:35:07.551005 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:07 crc kubenswrapper[4739]: I1201 15:35:07.551029 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:07 crc kubenswrapper[4739]: I1201 15:35:07.560636 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 01 15:35:07 crc kubenswrapper[4739]: I1201 15:35:07.560761 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:07 crc kubenswrapper[4739]: I1201 15:35:07.561854 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:07 crc kubenswrapper[4739]: I1201 15:35:07.561889 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:07 crc kubenswrapper[4739]: I1201 15:35:07.561899 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:08 crc kubenswrapper[4739]: E1201 15:35:08.550311 4739 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 15:35:09 crc kubenswrapper[4739]: I1201 15:35:09.468251 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 01 15:35:09 crc kubenswrapper[4739]: I1201 15:35:09.468601 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:09 crc kubenswrapper[4739]: I1201 15:35:09.470220 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:09 crc kubenswrapper[4739]: I1201 15:35:09.470274 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:09 crc kubenswrapper[4739]: I1201 15:35:09.470291 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:10 crc kubenswrapper[4739]: W1201 15:35:10.074311 4739 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 01 15:35:10 crc kubenswrapper[4739]: I1201 15:35:10.074506 4739 trace.go:236] Trace[1080412172]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 15:35:00.073) (total time: 10001ms): Dec 01 15:35:10 crc kubenswrapper[4739]: Trace[1080412172]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (15:35:10.074) Dec 01 15:35:10 crc kubenswrapper[4739]: Trace[1080412172]: [10.001225232s] [10.001225232s] END Dec 01 15:35:10 crc kubenswrapper[4739]: E1201 15:35:10.074549 4739 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 01 15:35:10 crc kubenswrapper[4739]: I1201 15:35:10.386498 4739 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 01 15:35:10 crc kubenswrapper[4739]: I1201 15:35:10.386585 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 01 15:35:10 crc kubenswrapper[4739]: I1201 15:35:10.406236 4739 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 01 15:35:10 crc kubenswrapper[4739]: W1201 15:35:10.950550 4739 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 01 15:35:10 crc kubenswrapper[4739]: I1201 15:35:10.950628 4739 trace.go:236] Trace[861595953]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 15:35:00.948) (total time: 10002ms): Dec 01 15:35:10 crc kubenswrapper[4739]: Trace[861595953]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (15:35:10.950) Dec 01 15:35:10 crc kubenswrapper[4739]: Trace[861595953]: [10.00253873s] [10.00253873s] END Dec 01 15:35:10 crc kubenswrapper[4739]: E1201 15:35:10.950646 4739 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 01 15:35:11 crc kubenswrapper[4739]: E1201 15:35:11.415387 4739 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 01 15:35:11 crc kubenswrapper[4739]: E1201 15:35:11.655884 4739 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 01 15:35:11 crc kubenswrapper[4739]: I1201 15:35:11.819523 4739 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 01 15:35:11 crc kubenswrapper[4739]: I1201 15:35:11.819789 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 01 15:35:11 crc kubenswrapper[4739]: I1201 15:35:11.828452 4739 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 01 15:35:11 crc kubenswrapper[4739]: I1201 15:35:11.828612 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 01 15:35:12 crc kubenswrapper[4739]: I1201 15:35:12.844235 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:35:12 crc kubenswrapper[4739]: I1201 15:35:12.844387 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:12 crc kubenswrapper[4739]: I1201 15:35:12.846077 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:12 crc kubenswrapper[4739]: I1201 15:35:12.846223 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:12 crc kubenswrapper[4739]: I1201 15:35:12.846336 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:12 crc kubenswrapper[4739]: I1201 15:35:12.902470 4739 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]log ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]etcd ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/generic-apiserver-start-informers ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/priority-and-fairness-filter ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/start-apiextensions-informers ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/start-apiextensions-controllers ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/crd-informer-synced ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/start-system-namespaces-controller ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 01 15:35:12 crc kubenswrapper[4739]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/bootstrap-controller ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/start-kube-aggregator-informers ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/apiservice-registration-controller ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/apiservice-discovery-controller ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]autoregister-completion ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/apiservice-openapi-controller ok Dec 01 15:35:12 crc kubenswrapper[4739]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 01 15:35:12 crc kubenswrapper[4739]: livez check failed Dec 01 15:35:12 crc kubenswrapper[4739]: I1201 15:35:12.903649 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:35:14 crc kubenswrapper[4739]: I1201 15:35:14.856751 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:14 crc kubenswrapper[4739]: I1201 15:35:14.858350 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:14 crc kubenswrapper[4739]: I1201 15:35:14.858450 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:14 crc kubenswrapper[4739]: I1201 15:35:14.858476 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:14 crc kubenswrapper[4739]: I1201 15:35:14.858522 4739 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 15:35:14 crc kubenswrapper[4739]: E1201 15:35:14.863523 4739 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.301676 4739 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.408234 4739 apiserver.go:52] "Watching apiserver" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.413225 4739 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.413656 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.414116 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.414282 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.414348 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:15 crc kubenswrapper[4739]: E1201 15:35:15.414489 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.414361 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:15 crc kubenswrapper[4739]: E1201 15:35:15.414604 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.414819 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:15 crc kubenswrapper[4739]: E1201 15:35:15.414921 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.415005 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.416775 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.417032 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.417402 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.417621 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.417655 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.418913 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.418977 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.419274 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.419461 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.456818 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.472458 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.490307 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.509021 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.511074 4739 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.524875 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.539238 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:15 crc kubenswrapper[4739]: I1201 15:35:15.551031 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.826903 4739 trace.go:236] Trace[292190899]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 15:35:01.880) (total time: 14946ms): Dec 01 15:35:16 crc kubenswrapper[4739]: Trace[292190899]: ---"Objects listed" error: 14946ms (15:35:16.826) Dec 01 15:35:16 crc kubenswrapper[4739]: Trace[292190899]: [14.946324682s] [14.946324682s] END Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.827484 4739 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.828122 4739 trace.go:236] Trace[1756667388]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 15:35:01.954) (total time: 14873ms): Dec 01 15:35:16 crc kubenswrapper[4739]: Trace[1756667388]: ---"Objects listed" error: 14872ms (15:35:16.827) Dec 01 15:35:16 crc kubenswrapper[4739]: Trace[1756667388]: [14.873091874s] [14.873091874s] END Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.828158 4739 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.828395 4739 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.862275 4739 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36826->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.862359 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36826->192.168.126.11:17697: read: connection reset by peer" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.929530 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.929575 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.929593 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.929609 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.929627 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.929646 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.929663 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.929684 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.929705 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.929730 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930045 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930076 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930098 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930119 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930118 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930142 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930209 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930259 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930284 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930329 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930366 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930402 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930456 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930484 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930511 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930537 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930558 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930579 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930599 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930622 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930644 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930668 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.931191 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.931256 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.931287 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.931327 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.931360 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930386 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930404 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.930530 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: E1201 15:35:16.931393 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:35:17.431358384 +0000 UTC m=+19.257104518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.933235 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.933975 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934015 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934034 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934168 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934227 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934252 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934266 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934333 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.931671 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934380 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934448 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934487 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934518 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934560 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934593 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934619 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935015 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935050 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935078 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935224 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935256 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935279 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935298 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935321 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935346 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935371 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935390 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935411 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935455 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.931984 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935476 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935473 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.932083 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.932073 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.933332 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.932117 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.932150 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.932290 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.932374 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.932507 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.932540 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.932760 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.932779 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.932827 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.932870 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.933115 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.933155 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.933615 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.933770 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.933836 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934637 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934682 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934745 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.934800 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935410 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935775 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.935500 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936038 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936089 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936020 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936140 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936200 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936242 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936290 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936341 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936377 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936447 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936511 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936573 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936614 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936659 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936704 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936742 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936785 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936824 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936862 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936909 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936952 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.938862 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.938915 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.939179 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.939273 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.939746 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.939784 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.939812 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.939845 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.939877 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.939910 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.939936 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.940028 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.940163 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.940192 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.940222 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936142 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936228 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936533 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936944 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.936959 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.937188 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.937271 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.937365 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.937583 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.938147 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.938223 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.938366 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.939197 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.939283 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.939635 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.939810 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.940187 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.940344 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.940375 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.940841 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.940881 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942152 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942215 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942327 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942369 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942406 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942512 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942559 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942606 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942648 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942694 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942744 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942789 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942838 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.940382 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942883 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.940880 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.941093 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942924 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.942963 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943001 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943041 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943077 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943118 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943162 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943199 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943239 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943280 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943319 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943371 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943412 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943472 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943512 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943552 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943592 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943625 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943664 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943705 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943800 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943843 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943881 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943916 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943960 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943999 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944040 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944075 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944117 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944160 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944195 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944246 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944300 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944355 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944409 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944500 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944562 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944615 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944674 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944735 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944789 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944852 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944908 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945051 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945102 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945164 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945221 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945273 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945334 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945391 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945470 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945552 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945610 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945667 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945720 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945780 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945865 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945919 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945981 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946038 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946100 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946150 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946213 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946289 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946346 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946404 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946534 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946597 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946656 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946712 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946768 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946828 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946887 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946991 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.947046 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.947103 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.947157 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.947211 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.947262 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.949149 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.949205 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.949320 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.949366 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.949444 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.951481 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.952381 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.941155 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.941909 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.943864 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.944618 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.945718 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946255 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946158 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946525 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946703 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946725 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.946920 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.947344 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.947482 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.952813 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.947574 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.947781 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.947946 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.948187 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.948344 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.948403 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.948487 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.949564 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.949542 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.949620 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.949646 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.949959 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.949987 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.950181 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.950643 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.950713 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.951140 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.951214 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.951232 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.951248 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.952021 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.953385 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.953501 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.955681 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.956117 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.956117 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.956346 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.956554 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.956856 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.956866 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.957089 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.957104 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.957274 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.957359 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.957889 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.957920 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.957861 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.957939 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.958003 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.958263 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.958512 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.958352 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.958739 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.958736 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.958775 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.958820 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.958827 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.959174 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.959190 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.959258 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.959274 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.959507 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.959588 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.959601 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.959691 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.960441 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.960947 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.960964 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.961340 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.961473 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.961603 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.961757 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.961797 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.962027 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.962029 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.962408 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.962479 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.962512 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.962518 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.963573 4739 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.963601 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.963833 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.964746 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.964784 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.964862 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.964906 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.965549 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.965732 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 15:35:16 crc kubenswrapper[4739]: E1201 15:35:16.965802 4739 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:35:16 crc kubenswrapper[4739]: E1201 15:35:16.965970 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:17.465947564 +0000 UTC m=+19.291693658 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.966263 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.966538 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.966815 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967003 4739 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967095 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967142 4739 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967239 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967256 4739 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967267 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967278 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967288 4739 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967297 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967307 4739 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967316 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967325 4739 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967335 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967345 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967355 4739 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967364 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967373 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967383 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967392 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967401 4739 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967410 4739 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967460 4739 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967469 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967478 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967488 4739 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967498 4739 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967507 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967517 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967526 4739 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967535 4739 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967543 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967552 4739 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967561 4739 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967570 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967580 4739 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967759 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.967108 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.968711 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.968902 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.968926 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.968940 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.968949 4739 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.968959 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.968969 4739 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.968979 4739 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.968991 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969000 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969011 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969020 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969030 4739 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969049 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969061 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969072 4739 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969089 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969101 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969110 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969120 4739 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969129 4739 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969138 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969148 4739 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969157 4739 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969168 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969180 4739 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969193 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969207 4739 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969219 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969232 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969245 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969257 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969267 4739 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969278 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969289 4739 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969300 4739 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969310 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969321 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.969803 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.972369 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.972727 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.973357 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.973663 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.974037 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.974724 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.979274 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.980065 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.980739 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.980806 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.980859 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.981143 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.981386 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.981468 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.981833 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.982030 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.982387 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.982527 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.982893 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.983470 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: E1201 15:35:16.983549 4739 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:35:16 crc kubenswrapper[4739]: E1201 15:35:16.983618 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:17.483595521 +0000 UTC m=+19.309341625 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.983825 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: E1201 15:35:16.984007 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:35:16 crc kubenswrapper[4739]: E1201 15:35:16.984024 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:35:16 crc kubenswrapper[4739]: E1201 15:35:16.984040 4739 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:16 crc kubenswrapper[4739]: E1201 15:35:16.984086 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:17.48407558 +0000 UTC m=+19.309821694 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:16 crc kubenswrapper[4739]: E1201 15:35:16.985228 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:35:16 crc kubenswrapper[4739]: E1201 15:35:16.985261 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:35:16 crc kubenswrapper[4739]: E1201 15:35:16.985279 4739 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:16 crc kubenswrapper[4739]: E1201 15:35:16.985393 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:17.485366875 +0000 UTC m=+19.311112979 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.986571 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.987335 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.987883 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.988130 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.988273 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.988813 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.988834 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.988951 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.989266 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.991267 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.993477 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.996261 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.996352 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.996447 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.996861 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.996808 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.996729 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.997045 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.997384 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.999023 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:16 crc kubenswrapper[4739]: I1201 15:35:16.999175 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.000935 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.002002 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.002009 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.002159 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.002375 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.002847 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.003090 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.004713 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.007258 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.012968 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.024147 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.027399 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070436 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070489 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070560 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070575 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070587 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070597 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070605 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070615 4739 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070623 4739 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070631 4739 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070640 4739 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070648 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070656 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070664 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070672 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070679 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070687 4739 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070695 4739 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070704 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070712 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070720 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070728 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070736 4739 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070745 4739 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070753 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070761 4739 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070769 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070779 4739 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070787 4739 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070796 4739 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070805 4739 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070814 4739 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070822 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070831 4739 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070840 4739 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070850 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070857 4739 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070865 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070877 4739 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070885 4739 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070894 4739 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070904 4739 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070914 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070924 4739 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070935 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070948 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070958 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070967 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070976 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070986 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.070996 4739 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071006 4739 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071015 4739 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071023 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071031 4739 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071039 4739 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071047 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071055 4739 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071064 4739 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071072 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071084 4739 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071082 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071114 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071092 4739 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071143 4739 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071163 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071198 4739 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071216 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071233 4739 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071248 4739 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071263 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071280 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071296 4739 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071311 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071327 4739 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071341 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071356 4739 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071371 4739 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071386 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071401 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071442 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071461 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071476 4739 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071492 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071509 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071527 4739 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071543 4739 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071559 4739 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071575 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071591 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071607 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071624 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071641 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071657 4739 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071674 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071689 4739 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071707 4739 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071724 4739 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071740 4739 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071758 4739 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071775 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071791 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071809 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071826 4739 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071843 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071859 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071877 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071893 4739 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071909 4739 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071925 4739 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071940 4739 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071956 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071972 4739 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.071989 4739 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.072005 4739 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.072021 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.072037 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.072059 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.072077 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.072094 4739 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.072110 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.072127 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.072144 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.072160 4739 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.072176 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.072192 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.072210 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.241589 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.252547 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.267017 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.408392 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.420851 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.423061 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.423492 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.439238 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.455701 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.471968 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.475345 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.475498 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.475564 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:35:18.475529814 +0000 UTC m=+20.301275908 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.475622 4739 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.475688 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:18.475672227 +0000 UTC m=+20.301418391 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.476124 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.476140 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.476273 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.476240 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.476366 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.476468 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.486181 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.498864 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.512452 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.526662 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.536029 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.552208 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.561668 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.571885 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.574681 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"617fce32eef9f94df39383de2e4bf97996ab800f76678aab6021fe8b12c8c1a2"} Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.575772 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.575816 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.575851 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.575906 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883"} Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.575933 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.575943 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f7ca8ed3b722311eaf58daf93ae37d4ebec2e35e4c2d371c90f8818c2a821387"} Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.575956 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.575973 4739 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.576017 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:18.576002902 +0000 UTC m=+20.401749006 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.576016 4739 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.576058 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:18.576046143 +0000 UTC m=+20.401792247 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.575933 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.576103 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.576114 4739 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:17 crc kubenswrapper[4739]: E1201 15:35:17.576139 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:18.576131065 +0000 UTC m=+20.401877159 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.578041 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4"} Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.578087 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324"} Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.578101 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4739f4c6c4da7f4d1dcdddaebc4fca9e1318206f0ae9dd8b97f62e0063e808be"} Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.580556 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.582875 4739 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4" exitCode=255 Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.582937 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4"} Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.585021 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.591433 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.591838 4739 scope.go:117] "RemoveContainer" containerID="ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.594462 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.610156 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.621599 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.636082 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.651566 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.669543 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.689013 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.903676 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.920677 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.936597 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.948647 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.962062 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.980655 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.992407 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:17Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:17 crc kubenswrapper[4739]: I1201 15:35:17.996992 4739 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.006950 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.018481 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.480138 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.481058 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.482037 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.482902 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.483761 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.484878 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:35:18 crc kubenswrapper[4739]: E1201 15:35:18.485043 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:35:20.485020639 +0000 UTC m=+22.310766743 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.485095 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:18 crc kubenswrapper[4739]: E1201 15:35:18.485200 4739 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:35:18 crc kubenswrapper[4739]: E1201 15:35:18.485244 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:20.485235213 +0000 UTC m=+22.310981307 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.485862 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.486860 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.487610 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.489028 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.489741 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.491014 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.491966 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.493142 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.493866 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.494600 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.495315 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.496029 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.496802 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.498028 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.498870 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.500220 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.500913 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.501780 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.503010 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.504006 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.507637 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.508773 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.511316 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.512896 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.513414 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.514499 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.515171 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.515948 4739 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.516103 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.518388 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.520965 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.521780 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.524370 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.524527 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.526123 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.526917 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.527965 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.528632 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.529733 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.530353 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.531311 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.531910 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.532771 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.533599 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.534675 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.535510 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.536460 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.536968 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.537861 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.538662 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.539261 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.540095 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.545951 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.568035 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.572342 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.581684 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.587115 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.587190 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.587244 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:18 crc kubenswrapper[4739]: E1201 15:35:18.587399 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:35:18 crc kubenswrapper[4739]: E1201 15:35:18.587442 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:35:18 crc kubenswrapper[4739]: E1201 15:35:18.587466 4739 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:18 crc kubenswrapper[4739]: E1201 15:35:18.587531 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:20.587503476 +0000 UTC m=+22.413249570 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:18 crc kubenswrapper[4739]: E1201 15:35:18.587543 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:35:18 crc kubenswrapper[4739]: E1201 15:35:18.587578 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:35:18 crc kubenswrapper[4739]: E1201 15:35:18.587593 4739 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:18 crc kubenswrapper[4739]: E1201 15:35:18.587634 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:20.587624487 +0000 UTC m=+22.413370581 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:18 crc kubenswrapper[4739]: E1201 15:35:18.587640 4739 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:35:18 crc kubenswrapper[4739]: E1201 15:35:18.587788 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:20.58774754 +0000 UTC m=+22.413493704 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.590182 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.593513 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41"} Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.593554 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.598450 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.603207 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.618144 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.654458 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.679089 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.700861 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.713388 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.727689 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.741738 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.755186 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:18 crc kubenswrapper[4739]: I1201 15:35:18.769667 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.070638 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-l4wkn"] Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.070947 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-vprlf"] Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.071145 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-l4wkn" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.071693 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-nt6tv"] Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.071904 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.071922 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.073873 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.075124 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.075175 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.076123 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.076189 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.076200 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.076270 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.076283 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.076508 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.076728 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.090286 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.101805 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.115729 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.126279 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.138372 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.155331 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.172518 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.188322 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.190619 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-os-release\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.190664 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhdxx\" (UniqueName: \"kubernetes.io/projected/341a0b94-e13d-45cb-8f1c-2c1d1f101272-kube-api-access-mhdxx\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.190692 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-cnibin\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.190776 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a46c661b-7815-415e-90ac-4be28c3da8bf-system-cni-dir\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.190819 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8bwz\" (UniqueName: \"kubernetes.io/projected/a46c661b-7815-415e-90ac-4be28c3da8bf-kube-api-access-k8bwz\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.190852 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-hostroot\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.190885 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/341a0b94-e13d-45cb-8f1c-2c1d1f101272-multus-daemon-config\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.190953 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a46c661b-7815-415e-90ac-4be28c3da8bf-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.190982 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8ssk\" (UniqueName: \"kubernetes.io/projected/00331ca6-51ac-457e-b95c-9841bda7b582-kube-api-access-l8ssk\") pod \"node-resolver-l4wkn\" (UID: \"00331ca6-51ac-457e-b95c-9841bda7b582\") " pod="openshift-dns/node-resolver-l4wkn" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191013 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-run-multus-certs\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191043 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-system-cni-dir\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191098 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a46c661b-7815-415e-90ac-4be28c3da8bf-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191128 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-var-lib-cni-bin\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191157 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-var-lib-kubelet\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191186 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a46c661b-7815-415e-90ac-4be28c3da8bf-cnibin\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191217 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a46c661b-7815-415e-90ac-4be28c3da8bf-os-release\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191262 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-run-k8s-cni-cncf-io\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191290 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-multus-conf-dir\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191319 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-multus-socket-dir-parent\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191345 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-etc-kubernetes\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191373 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-multus-cni-dir\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191401 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/341a0b94-e13d-45cb-8f1c-2c1d1f101272-cni-binary-copy\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191457 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-var-lib-cni-multus\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191489 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/00331ca6-51ac-457e-b95c-9841bda7b582-hosts-file\") pod \"node-resolver-l4wkn\" (UID: \"00331ca6-51ac-457e-b95c-9841bda7b582\") " pod="openshift-dns/node-resolver-l4wkn" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191516 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-run-netns\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.191568 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a46c661b-7815-415e-90ac-4be28c3da8bf-cni-binary-copy\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.201190 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.219725 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.235228 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.246287 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.262724 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.277365 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292165 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a46c661b-7815-415e-90ac-4be28c3da8bf-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292445 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8ssk\" (UniqueName: \"kubernetes.io/projected/00331ca6-51ac-457e-b95c-9841bda7b582-kube-api-access-l8ssk\") pod \"node-resolver-l4wkn\" (UID: \"00331ca6-51ac-457e-b95c-9841bda7b582\") " pod="openshift-dns/node-resolver-l4wkn" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292533 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-run-multus-certs\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292619 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-run-multus-certs\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292629 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-system-cni-dir\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292688 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-var-lib-kubelet\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292708 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a46c661b-7815-415e-90ac-4be28c3da8bf-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292725 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-var-lib-cni-bin\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292743 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a46c661b-7815-415e-90ac-4be28c3da8bf-cnibin\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292759 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a46c661b-7815-415e-90ac-4be28c3da8bf-os-release\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292790 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-run-k8s-cni-cncf-io\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292810 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-multus-conf-dir\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292788 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a46c661b-7815-415e-90ac-4be28c3da8bf-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292859 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a46c661b-7815-415e-90ac-4be28c3da8bf-cnibin\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292827 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/341a0b94-e13d-45cb-8f1c-2c1d1f101272-cni-binary-copy\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292933 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-multus-socket-dir-parent\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292939 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-run-k8s-cni-cncf-io\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292948 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-multus-conf-dir\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292995 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-etc-kubernetes\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292969 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-etc-kubernetes\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293034 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-multus-cni-dir\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292948 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-var-lib-kubelet\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.292948 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-var-lib-cni-bin\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293034 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-multus-socket-dir-parent\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293081 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-run-netns\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293059 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-run-netns\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293158 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-var-lib-cni-multus\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293185 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/00331ca6-51ac-457e-b95c-9841bda7b582-hosts-file\") pod \"node-resolver-l4wkn\" (UID: \"00331ca6-51ac-457e-b95c-9841bda7b582\") " pod="openshift-dns/node-resolver-l4wkn" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293192 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a46c661b-7815-415e-90ac-4be28c3da8bf-os-release\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293205 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a46c661b-7815-415e-90ac-4be28c3da8bf-cni-binary-copy\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293246 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-host-var-lib-cni-multus\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293260 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/00331ca6-51ac-457e-b95c-9841bda7b582-hosts-file\") pod \"node-resolver-l4wkn\" (UID: \"00331ca6-51ac-457e-b95c-9841bda7b582\") " pod="openshift-dns/node-resolver-l4wkn" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293297 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-os-release\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293324 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhdxx\" (UniqueName: \"kubernetes.io/projected/341a0b94-e13d-45cb-8f1c-2c1d1f101272-kube-api-access-mhdxx\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293364 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-cnibin\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293397 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a46c661b-7815-415e-90ac-4be28c3da8bf-system-cni-dir\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293323 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-multus-cni-dir\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293441 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8bwz\" (UniqueName: \"kubernetes.io/projected/a46c661b-7815-415e-90ac-4be28c3da8bf-kube-api-access-k8bwz\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293531 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-hostroot\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293557 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/341a0b94-e13d-45cb-8f1c-2c1d1f101272-multus-daemon-config\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293592 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-cnibin\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293634 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a46c661b-7815-415e-90ac-4be28c3da8bf-system-cni-dir\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293400 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-os-release\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293648 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/341a0b94-e13d-45cb-8f1c-2c1d1f101272-cni-binary-copy\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293645 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-hostroot\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293661 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a46c661b-7815-415e-90ac-4be28c3da8bf-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293912 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/341a0b94-e13d-45cb-8f1c-2c1d1f101272-system-cni-dir\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.293923 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a46c661b-7815-415e-90ac-4be28c3da8bf-cni-binary-copy\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.294104 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.294305 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/341a0b94-e13d-45cb-8f1c-2c1d1f101272-multus-daemon-config\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.308112 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.311370 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8bwz\" (UniqueName: \"kubernetes.io/projected/a46c661b-7815-415e-90ac-4be28c3da8bf-kube-api-access-k8bwz\") pod \"multus-additional-cni-plugins-vprlf\" (UID: \"a46c661b-7815-415e-90ac-4be28c3da8bf\") " pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.313545 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8ssk\" (UniqueName: \"kubernetes.io/projected/00331ca6-51ac-457e-b95c-9841bda7b582-kube-api-access-l8ssk\") pod \"node-resolver-l4wkn\" (UID: \"00331ca6-51ac-457e-b95c-9841bda7b582\") " pod="openshift-dns/node-resolver-l4wkn" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.317203 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhdxx\" (UniqueName: \"kubernetes.io/projected/341a0b94-e13d-45cb-8f1c-2c1d1f101272-kube-api-access-mhdxx\") pod \"multus-nt6tv\" (UID: \"341a0b94-e13d-45cb-8f1c-2c1d1f101272\") " pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.327855 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.340547 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.353557 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.369169 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.383252 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-l4wkn" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.390347 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-nt6tv" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.395377 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vprlf" Dec 01 15:35:19 crc kubenswrapper[4739]: W1201 15:35:19.402437 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00331ca6_51ac_457e_b95c_9841bda7b582.slice/crio-4d34866105ed6395c048b65a4f720bc2b5a7a212a7730ea10f3614e346009357 WatchSource:0}: Error finding container 4d34866105ed6395c048b65a4f720bc2b5a7a212a7730ea10f3614e346009357: Status 404 returned error can't find the container with id 4d34866105ed6395c048b65a4f720bc2b5a7a212a7730ea10f3614e346009357 Dec 01 15:35:19 crc kubenswrapper[4739]: W1201 15:35:19.411370 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod341a0b94_e13d_45cb_8f1c_2c1d1f101272.slice/crio-6ab347af22baa41df57bceab58c3615114bd209e6663f7a5cf22fb2c138271ad WatchSource:0}: Error finding container 6ab347af22baa41df57bceab58c3615114bd209e6663f7a5cf22fb2c138271ad: Status 404 returned error can't find the container with id 6ab347af22baa41df57bceab58c3615114bd209e6663f7a5cf22fb2c138271ad Dec 01 15:35:19 crc kubenswrapper[4739]: W1201 15:35:19.412989 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda46c661b_7815_415e_90ac_4be28c3da8bf.slice/crio-b565d91f3062dfcf9c5259b496a200b95a9aeee2077514f7b779f6be0fc5e3f7 WatchSource:0}: Error finding container b565d91f3062dfcf9c5259b496a200b95a9aeee2077514f7b779f6be0fc5e3f7: Status 404 returned error can't find the container with id b565d91f3062dfcf9c5259b496a200b95a9aeee2077514f7b779f6be0fc5e3f7 Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.441375 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dlvkg"] Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.442040 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-rnmn7"] Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.442174 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.442378 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.444390 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.444454 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.444742 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.445348 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.445492 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.445749 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.445775 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.446875 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.448335 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.448655 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.449192 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.449906 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.473072 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.478041 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.478110 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:19 crc kubenswrapper[4739]: E1201 15:35:19.478201 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.478067 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:19 crc kubenswrapper[4739]: E1201 15:35:19.478303 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:19 crc kubenswrapper[4739]: E1201 15:35:19.478510 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.494129 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.501519 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.507084 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.512584 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.520064 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.525212 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.544001 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.556265 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.570094 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.592914 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595544 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-systemd\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595585 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovnkube-config\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595613 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvdpq\" (UniqueName: \"kubernetes.io/projected/43a8da1c-7c68-4af9-8b57-817e55d3b875-kube-api-access-rvdpq\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595636 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-env-overrides\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595656 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e-mcd-auth-proxy-config\") pod \"machine-config-daemon-rnmn7\" (UID: \"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\") " pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595686 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-cni-bin\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595705 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovn-node-metrics-cert\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595724 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovnkube-script-lib\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595744 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bklr6\" (UniqueName: \"kubernetes.io/projected/dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e-kube-api-access-bklr6\") pod \"machine-config-daemon-rnmn7\" (UID: \"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\") " pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595772 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-cni-netd\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595787 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e-rootfs\") pod \"machine-config-daemon-rnmn7\" (UID: \"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\") " pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595802 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-slash\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595816 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-run-netns\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595830 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-openvswitch\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595844 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-systemd-units\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595858 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-node-log\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595876 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-var-lib-openvswitch\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595897 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-log-socket\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595919 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-run-ovn-kubernetes\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595938 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595958 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-kubelet\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.595994 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-etc-openvswitch\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.596157 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e-proxy-tls\") pod \"machine-config-daemon-rnmn7\" (UID: \"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\") " pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.596346 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-ovn\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.598162 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-l4wkn" event={"ID":"00331ca6-51ac-457e-b95c-9841bda7b582","Type":"ContainerStarted","Data":"4d34866105ed6395c048b65a4f720bc2b5a7a212a7730ea10f3614e346009357"} Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.600865 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" event={"ID":"a46c661b-7815-415e-90ac-4be28c3da8bf","Type":"ContainerStarted","Data":"b565d91f3062dfcf9c5259b496a200b95a9aeee2077514f7b779f6be0fc5e3f7"} Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.604874 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nt6tv" event={"ID":"341a0b94-e13d-45cb-8f1c-2c1d1f101272","Type":"ContainerStarted","Data":"13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46"} Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.604940 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nt6tv" event={"ID":"341a0b94-e13d-45cb-8f1c-2c1d1f101272","Type":"ContainerStarted","Data":"6ab347af22baa41df57bceab58c3615114bd209e6663f7a5cf22fb2c138271ad"} Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.610503 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.639784 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.667995 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.696203 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697526 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-slash\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697571 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-run-netns\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697594 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-openvswitch\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697617 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-cni-netd\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697647 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e-rootfs\") pod \"machine-config-daemon-rnmn7\" (UID: \"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\") " pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697668 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-cni-netd\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697686 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-systemd-units\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697708 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-node-log\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697706 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-openvswitch\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697728 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-node-log\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697648 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-run-netns\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697656 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-slash\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697752 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e-rootfs\") pod \"machine-config-daemon-rnmn7\" (UID: \"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\") " pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697711 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-systemd-units\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697768 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-log-socket\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697729 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-log-socket\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697886 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-run-ovn-kubernetes\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697914 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697934 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-run-ovn-kubernetes\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697954 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-var-lib-openvswitch\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697988 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-var-lib-openvswitch\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697993 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-kubelet\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698017 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-kubelet\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698049 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-etc-openvswitch\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698074 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e-proxy-tls\") pod \"machine-config-daemon-rnmn7\" (UID: \"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\") " pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.697992 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698105 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-etc-openvswitch\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698126 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-ovn\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698149 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-systemd\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698182 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovnkube-config\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698193 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-ovn\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698205 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvdpq\" (UniqueName: \"kubernetes.io/projected/43a8da1c-7c68-4af9-8b57-817e55d3b875-kube-api-access-rvdpq\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698230 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-env-overrides\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698251 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e-mcd-auth-proxy-config\") pod \"machine-config-daemon-rnmn7\" (UID: \"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\") " pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698274 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-cni-bin\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698303 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovn-node-metrics-cert\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698324 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovnkube-script-lib\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698410 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bklr6\" (UniqueName: \"kubernetes.io/projected/dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e-kube-api-access-bklr6\") pod \"machine-config-daemon-rnmn7\" (UID: \"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\") " pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698552 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-cni-bin\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698251 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-systemd\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698863 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-env-overrides\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698936 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovnkube-config\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.698961 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e-mcd-auth-proxy-config\") pod \"machine-config-daemon-rnmn7\" (UID: \"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\") " pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.699161 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovnkube-script-lib\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.700919 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e-proxy-tls\") pod \"machine-config-daemon-rnmn7\" (UID: \"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\") " pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.701140 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovn-node-metrics-cert\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.730025 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bklr6\" (UniqueName: \"kubernetes.io/projected/dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e-kube-api-access-bklr6\") pod \"machine-config-daemon-rnmn7\" (UID: \"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\") " pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.733136 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvdpq\" (UniqueName: \"kubernetes.io/projected/43a8da1c-7c68-4af9-8b57-817e55d3b875-kube-api-access-rvdpq\") pod \"ovnkube-node-dlvkg\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.738947 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.757974 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.764606 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.771812 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.774354 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.786514 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.804875 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.815777 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.828689 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.845545 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.868715 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.883915 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.896143 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.908342 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.924898 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.935779 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:19 crc kubenswrapper[4739]: I1201 15:35:19.949738 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.506673 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:35:20 crc kubenswrapper[4739]: E1201 15:35:20.506994 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:35:24.506956355 +0000 UTC m=+26.332702489 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.507327 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:20 crc kubenswrapper[4739]: E1201 15:35:20.507548 4739 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:35:20 crc kubenswrapper[4739]: E1201 15:35:20.507679 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:24.507649839 +0000 UTC m=+26.333395983 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.608809 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.608879 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.608931 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:20 crc kubenswrapper[4739]: E1201 15:35:20.609044 4739 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:35:20 crc kubenswrapper[4739]: E1201 15:35:20.609094 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:35:20 crc kubenswrapper[4739]: E1201 15:35:20.609118 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:35:20 crc kubenswrapper[4739]: E1201 15:35:20.609125 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:24.609100385 +0000 UTC m=+26.434846519 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:35:20 crc kubenswrapper[4739]: E1201 15:35:20.609137 4739 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:20 crc kubenswrapper[4739]: E1201 15:35:20.609140 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:35:20 crc kubenswrapper[4739]: E1201 15:35:20.609201 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:24.609180407 +0000 UTC m=+26.434926531 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:20 crc kubenswrapper[4739]: E1201 15:35:20.609205 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:35:20 crc kubenswrapper[4739]: E1201 15:35:20.609239 4739 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:20 crc kubenswrapper[4739]: E1201 15:35:20.609358 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:24.609323419 +0000 UTC m=+26.435069563 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.610706 4739 generic.go:334] "Generic (PLEG): container finished" podID="a46c661b-7815-415e-90ac-4be28c3da8bf" containerID="2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a" exitCode=0 Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.610810 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" event={"ID":"a46c661b-7815-415e-90ac-4be28c3da8bf","Type":"ContainerDied","Data":"2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a"} Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.612452 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-l4wkn" event={"ID":"00331ca6-51ac-457e-b95c-9841bda7b582","Type":"ContainerStarted","Data":"61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426"} Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.613750 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e"} Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.616343 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df"} Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.616370 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e"} Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.616386 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"1a15e69ac81e6c25a481705cdd73e79a20214438b4882a110e77df2aa497c6ff"} Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.618549 4739 generic.go:334] "Generic (PLEG): container finished" podID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerID="e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b" exitCode=0 Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.618604 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerDied","Data":"e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b"} Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.618721 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerStarted","Data":"8f334e545ef5edead52a18e9b43acc2e71e4c4867917134a24881d8668f52997"} Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.638756 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.660367 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.676256 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.700283 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.711866 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.727270 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.744632 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.757796 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.786096 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.799843 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.814117 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.826582 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.843984 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.857894 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.879555 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.893878 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.907169 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.919890 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.933323 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.946022 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.960858 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.977097 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:20 crc kubenswrapper[4739]: I1201 15:35:20.992283 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.006242 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.021448 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.035987 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.053632 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.065852 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.264020 4739 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.266268 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.266323 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.266343 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.266494 4739 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.275986 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-5jnk6"] Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.276331 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5jnk6" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.280271 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.281689 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.282434 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.289256 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.293694 4739 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.294080 4739 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.295181 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.295210 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.295222 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.295240 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.295253 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:21Z","lastTransitionTime":"2025-12-01T15:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.313081 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.352664 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: E1201 15:35:21.360279 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.365546 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.365578 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.365587 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.365605 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.365615 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:21Z","lastTransitionTime":"2025-12-01T15:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.383963 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: E1201 15:35:21.388269 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.392141 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.392172 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.392180 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.392196 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.392207 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:21Z","lastTransitionTime":"2025-12-01T15:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.397319 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: E1201 15:35:21.409279 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.412793 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.412833 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.412843 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.412858 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.412870 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:21Z","lastTransitionTime":"2025-12-01T15:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.417697 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4mjh\" (UniqueName: \"kubernetes.io/projected/3613c34f-77d5-414c-b3ae-c31e123ad84b-kube-api-access-n4mjh\") pod \"node-ca-5jnk6\" (UID: \"3613c34f-77d5-414c-b3ae-c31e123ad84b\") " pod="openshift-image-registry/node-ca-5jnk6" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.417746 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.417782 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3613c34f-77d5-414c-b3ae-c31e123ad84b-host\") pod \"node-ca-5jnk6\" (UID: \"3613c34f-77d5-414c-b3ae-c31e123ad84b\") " pod="openshift-image-registry/node-ca-5jnk6" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.417861 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3613c34f-77d5-414c-b3ae-c31e123ad84b-serviceca\") pod \"node-ca-5jnk6\" (UID: \"3613c34f-77d5-414c-b3ae-c31e123ad84b\") " pod="openshift-image-registry/node-ca-5jnk6" Dec 01 15:35:21 crc kubenswrapper[4739]: E1201 15:35:21.423946 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.428484 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.428612 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.428713 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.428809 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.428911 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:21Z","lastTransitionTime":"2025-12-01T15:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.431974 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: E1201 15:35:21.440275 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: E1201 15:35:21.440627 4739 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.442046 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.442079 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.442090 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.442103 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.442114 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:21Z","lastTransitionTime":"2025-12-01T15:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.451527 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.465014 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.474937 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.476094 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.476105 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.476097 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:21 crc kubenswrapper[4739]: E1201 15:35:21.476207 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:21 crc kubenswrapper[4739]: E1201 15:35:21.476299 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:21 crc kubenswrapper[4739]: E1201 15:35:21.476375 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.488873 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.499760 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.510749 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.519390 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4mjh\" (UniqueName: \"kubernetes.io/projected/3613c34f-77d5-414c-b3ae-c31e123ad84b-kube-api-access-n4mjh\") pod \"node-ca-5jnk6\" (UID: \"3613c34f-77d5-414c-b3ae-c31e123ad84b\") " pod="openshift-image-registry/node-ca-5jnk6" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.519484 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3613c34f-77d5-414c-b3ae-c31e123ad84b-host\") pod \"node-ca-5jnk6\" (UID: \"3613c34f-77d5-414c-b3ae-c31e123ad84b\") " pod="openshift-image-registry/node-ca-5jnk6" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.519864 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3613c34f-77d5-414c-b3ae-c31e123ad84b-serviceca\") pod \"node-ca-5jnk6\" (UID: \"3613c34f-77d5-414c-b3ae-c31e123ad84b\") " pod="openshift-image-registry/node-ca-5jnk6" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.519734 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3613c34f-77d5-414c-b3ae-c31e123ad84b-host\") pod \"node-ca-5jnk6\" (UID: \"3613c34f-77d5-414c-b3ae-c31e123ad84b\") " pod="openshift-image-registry/node-ca-5jnk6" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.520963 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3613c34f-77d5-414c-b3ae-c31e123ad84b-serviceca\") pod \"node-ca-5jnk6\" (UID: \"3613c34f-77d5-414c-b3ae-c31e123ad84b\") " pod="openshift-image-registry/node-ca-5jnk6" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.527882 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.540822 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.543812 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4mjh\" (UniqueName: \"kubernetes.io/projected/3613c34f-77d5-414c-b3ae-c31e123ad84b-kube-api-access-n4mjh\") pod \"node-ca-5jnk6\" (UID: \"3613c34f-77d5-414c-b3ae-c31e123ad84b\") " pod="openshift-image-registry/node-ca-5jnk6" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.544593 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.544625 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.544636 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.544650 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.544658 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:21Z","lastTransitionTime":"2025-12-01T15:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.551584 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.605992 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5jnk6" Dec 01 15:35:21 crc kubenswrapper[4739]: W1201 15:35:21.620815 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3613c34f_77d5_414c_b3ae_c31e123ad84b.slice/crio-f942e810a66d2e6128be81962d5e5bc40ed6117612f2caa7b59e409ecc93b102 WatchSource:0}: Error finding container f942e810a66d2e6128be81962d5e5bc40ed6117612f2caa7b59e409ecc93b102: Status 404 returned error can't find the container with id f942e810a66d2e6128be81962d5e5bc40ed6117612f2caa7b59e409ecc93b102 Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.624847 4739 generic.go:334] "Generic (PLEG): container finished" podID="a46c661b-7815-415e-90ac-4be28c3da8bf" containerID="3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857" exitCode=0 Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.624860 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" event={"ID":"a46c661b-7815-415e-90ac-4be28c3da8bf","Type":"ContainerDied","Data":"3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.630258 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerStarted","Data":"63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.630291 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerStarted","Data":"d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.630307 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerStarted","Data":"9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.630319 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerStarted","Data":"bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.630332 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerStarted","Data":"a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.630343 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerStarted","Data":"76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.646850 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.646920 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.646936 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.646954 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.646966 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:21Z","lastTransitionTime":"2025-12-01T15:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.667341 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.681396 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.695667 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.710664 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.741694 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.751942 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.751973 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.751982 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.751995 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.752007 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:21Z","lastTransitionTime":"2025-12-01T15:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.788198 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.816405 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.854976 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.855012 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.855023 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.855039 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.855053 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:21Z","lastTransitionTime":"2025-12-01T15:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.859858 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.898903 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.942671 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.957533 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.957572 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.957597 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.957612 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.957624 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:21Z","lastTransitionTime":"2025-12-01T15:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:21 crc kubenswrapper[4739]: I1201 15:35:21.982312 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.024764 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.060148 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.060199 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.060213 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.060234 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.060249 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:22Z","lastTransitionTime":"2025-12-01T15:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.061664 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.106954 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.145123 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.163886 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.163936 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.163954 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.163982 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.164002 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:22Z","lastTransitionTime":"2025-12-01T15:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.266159 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.266194 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.266203 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.266217 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.266226 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:22Z","lastTransitionTime":"2025-12-01T15:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.369063 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.369120 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.369137 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.369161 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.369178 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:22Z","lastTransitionTime":"2025-12-01T15:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.472222 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.472270 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.472290 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.472328 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.472347 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:22Z","lastTransitionTime":"2025-12-01T15:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.574939 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.575018 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.575041 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.575071 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.575095 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:22Z","lastTransitionTime":"2025-12-01T15:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.636226 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5jnk6" event={"ID":"3613c34f-77d5-414c-b3ae-c31e123ad84b","Type":"ContainerStarted","Data":"cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6"} Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.636281 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5jnk6" event={"ID":"3613c34f-77d5-414c-b3ae-c31e123ad84b","Type":"ContainerStarted","Data":"f942e810a66d2e6128be81962d5e5bc40ed6117612f2caa7b59e409ecc93b102"} Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.639375 4739 generic.go:334] "Generic (PLEG): container finished" podID="a46c661b-7815-415e-90ac-4be28c3da8bf" containerID="9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273" exitCode=0 Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.639484 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" event={"ID":"a46c661b-7815-415e-90ac-4be28c3da8bf","Type":"ContainerDied","Data":"9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273"} Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.660059 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.676792 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.678006 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.678062 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.678089 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.678119 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.678140 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:22Z","lastTransitionTime":"2025-12-01T15:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.695297 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.714329 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.728619 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.741223 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.758878 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.778520 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.781047 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.781093 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.781109 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.781129 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.781146 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:22Z","lastTransitionTime":"2025-12-01T15:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.793826 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.808410 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.820295 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.832577 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.854211 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.869406 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.883612 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.883644 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.883652 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.883666 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.883675 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:22Z","lastTransitionTime":"2025-12-01T15:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.887451 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.906443 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.917869 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.932368 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.951556 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.969470 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.985929 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.985975 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.985986 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.986003 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.986014 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:22Z","lastTransitionTime":"2025-12-01T15:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:22 crc kubenswrapper[4739]: I1201 15:35:22.995349 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:22Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.018598 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.056956 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.087724 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.087755 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.087764 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.087777 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.087786 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:23Z","lastTransitionTime":"2025-12-01T15:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.099550 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.138150 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.181617 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.189912 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.189942 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.189951 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.189965 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.189975 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:23Z","lastTransitionTime":"2025-12-01T15:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.224904 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.258258 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.292119 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.292159 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.292170 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.292187 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.292198 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:23Z","lastTransitionTime":"2025-12-01T15:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.302605 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.340324 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.394036 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.394077 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.394094 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.394116 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.394130 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:23Z","lastTransitionTime":"2025-12-01T15:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.476216 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.476222 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:23 crc kubenswrapper[4739]: E1201 15:35:23.476496 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.476223 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:23 crc kubenswrapper[4739]: E1201 15:35:23.476608 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:23 crc kubenswrapper[4739]: E1201 15:35:23.476736 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.497063 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.497147 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.497170 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.497199 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.497222 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:23Z","lastTransitionTime":"2025-12-01T15:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.600695 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.600756 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.600778 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.600820 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.600845 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:23Z","lastTransitionTime":"2025-12-01T15:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.645577 4739 generic.go:334] "Generic (PLEG): container finished" podID="a46c661b-7815-415e-90ac-4be28c3da8bf" containerID="7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8" exitCode=0 Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.645645 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" event={"ID":"a46c661b-7815-415e-90ac-4be28c3da8bf","Type":"ContainerDied","Data":"7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8"} Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.672265 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.691607 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.704225 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.704268 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.704279 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.704295 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.704306 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:23Z","lastTransitionTime":"2025-12-01T15:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.717277 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.737990 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.758711 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.779921 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.793866 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.806189 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.806222 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.806230 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.806244 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.806254 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:23Z","lastTransitionTime":"2025-12-01T15:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.811205 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.835789 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.848134 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.860654 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.886836 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.904864 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.907962 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.907997 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.908005 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.908022 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.908032 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:23Z","lastTransitionTime":"2025-12-01T15:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.917652 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:23 crc kubenswrapper[4739]: I1201 15:35:23.940543 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.010600 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.010649 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.010662 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.010681 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.010694 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:24Z","lastTransitionTime":"2025-12-01T15:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.114516 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.114615 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.114638 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.114663 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.114680 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:24Z","lastTransitionTime":"2025-12-01T15:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.217530 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.217617 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.217636 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.217663 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.217680 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:24Z","lastTransitionTime":"2025-12-01T15:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.320389 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.320500 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.320526 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.320557 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.320579 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:24Z","lastTransitionTime":"2025-12-01T15:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.423054 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.423126 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.423149 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.423179 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.423201 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:24Z","lastTransitionTime":"2025-12-01T15:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.525706 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.525778 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.525806 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.525834 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.525856 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:24Z","lastTransitionTime":"2025-12-01T15:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.549236 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.549397 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:24 crc kubenswrapper[4739]: E1201 15:35:24.549452 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:35:32.54938758 +0000 UTC m=+34.375133724 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:35:24 crc kubenswrapper[4739]: E1201 15:35:24.549563 4739 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:35:24 crc kubenswrapper[4739]: E1201 15:35:24.549622 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:32.549604834 +0000 UTC m=+34.375350958 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.629349 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.629391 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.629403 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.629435 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.629448 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:24Z","lastTransitionTime":"2025-12-01T15:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.650159 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.650226 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.650274 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:24 crc kubenswrapper[4739]: E1201 15:35:24.650383 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:35:24 crc kubenswrapper[4739]: E1201 15:35:24.650479 4739 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:35:24 crc kubenswrapper[4739]: E1201 15:35:24.650388 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:35:24 crc kubenswrapper[4739]: E1201 15:35:24.650547 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:35:24 crc kubenswrapper[4739]: E1201 15:35:24.650566 4739 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:24 crc kubenswrapper[4739]: E1201 15:35:24.650596 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:32.650561363 +0000 UTC m=+34.476307497 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:35:24 crc kubenswrapper[4739]: E1201 15:35:24.650630 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:32.650613014 +0000 UTC m=+34.476359218 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:24 crc kubenswrapper[4739]: E1201 15:35:24.650412 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:35:24 crc kubenswrapper[4739]: E1201 15:35:24.650947 4739 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:24 crc kubenswrapper[4739]: E1201 15:35:24.651008 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:32.650987916 +0000 UTC m=+34.476734110 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.655086 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerStarted","Data":"13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f"} Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.659023 4739 generic.go:334] "Generic (PLEG): container finished" podID="a46c661b-7815-415e-90ac-4be28c3da8bf" containerID="d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc" exitCode=0 Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.659063 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" event={"ID":"a46c661b-7815-415e-90ac-4be28c3da8bf","Type":"ContainerDied","Data":"d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc"} Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.677374 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:24Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.694163 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:24Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.715170 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:24Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.731598 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.731658 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.731673 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.731691 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.731703 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:24Z","lastTransitionTime":"2025-12-01T15:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.737586 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:24Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.778225 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:24Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.801323 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:24Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.820944 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:24Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.835142 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.835172 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.835182 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.835196 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.835207 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:24Z","lastTransitionTime":"2025-12-01T15:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.838003 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:24Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.856344 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:24Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.868028 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:24Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.876963 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:24Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.888315 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:24Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.899733 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:24Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.908093 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:24Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.920437 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:24Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.937127 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.937172 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.937181 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.937196 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:24 crc kubenswrapper[4739]: I1201 15:35:24.937205 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:24Z","lastTransitionTime":"2025-12-01T15:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.040607 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.040668 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.040687 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.040712 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.040730 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:25Z","lastTransitionTime":"2025-12-01T15:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.144343 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.144409 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.144479 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.144514 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.144539 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:25Z","lastTransitionTime":"2025-12-01T15:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.247349 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.247471 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.247501 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.247531 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.247549 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:25Z","lastTransitionTime":"2025-12-01T15:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.350371 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.350475 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.350499 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.350529 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.350551 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:25Z","lastTransitionTime":"2025-12-01T15:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.454462 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.454544 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.454568 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.454597 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.454622 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:25Z","lastTransitionTime":"2025-12-01T15:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.477078 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.477213 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:25 crc kubenswrapper[4739]: E1201 15:35:25.477331 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.477096 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:25 crc kubenswrapper[4739]: E1201 15:35:25.477451 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:25 crc kubenswrapper[4739]: E1201 15:35:25.477541 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.556915 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.556973 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.556990 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.557013 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.557031 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:25Z","lastTransitionTime":"2025-12-01T15:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.659521 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.659583 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.659601 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.659627 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.659652 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:25Z","lastTransitionTime":"2025-12-01T15:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.666084 4739 generic.go:334] "Generic (PLEG): container finished" podID="a46c661b-7815-415e-90ac-4be28c3da8bf" containerID="70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0" exitCode=0 Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.666120 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" event={"ID":"a46c661b-7815-415e-90ac-4be28c3da8bf","Type":"ContainerDied","Data":"70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0"} Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.700678 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.720755 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.737205 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.754878 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.763377 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.763450 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.763469 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.763499 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.763515 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:25Z","lastTransitionTime":"2025-12-01T15:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.770187 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.782779 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.796817 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.809976 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.822087 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.835635 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.850972 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.865696 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.866285 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.866364 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.866390 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.866461 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.866491 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:25Z","lastTransitionTime":"2025-12-01T15:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.886949 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.897832 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.908086 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:25Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.969098 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.969133 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.969143 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.969157 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:25 crc kubenswrapper[4739]: I1201 15:35:25.969165 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:25Z","lastTransitionTime":"2025-12-01T15:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.072031 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.072106 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.072124 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.072152 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.072172 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:26Z","lastTransitionTime":"2025-12-01T15:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.174925 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.175240 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.175252 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.175271 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.175289 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:26Z","lastTransitionTime":"2025-12-01T15:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.278165 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.278217 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.278233 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.278256 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.278272 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:26Z","lastTransitionTime":"2025-12-01T15:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.381480 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.381530 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.381541 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.381557 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.381567 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:26Z","lastTransitionTime":"2025-12-01T15:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.483369 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.483409 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.483436 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.483449 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.483460 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:26Z","lastTransitionTime":"2025-12-01T15:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.586371 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.586455 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.586470 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.586490 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.586502 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:26Z","lastTransitionTime":"2025-12-01T15:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.679869 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" event={"ID":"a46c661b-7815-415e-90ac-4be28c3da8bf","Type":"ContainerStarted","Data":"0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343"} Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.684741 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerStarted","Data":"7201f18006d89a4fca89142fe792cc3def0fa026eae9d587af724f777a746805"} Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.685471 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.689534 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.689593 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.689614 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.689638 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.689658 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:26Z","lastTransitionTime":"2025-12-01T15:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.699309 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.716242 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.723714 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.738325 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.762520 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.779995 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.792642 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.792697 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.792713 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.792730 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.792742 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:26Z","lastTransitionTime":"2025-12-01T15:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.796773 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.815158 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.831467 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.846940 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.874906 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.893517 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.895471 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.895507 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.895520 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.895541 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.895553 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:26Z","lastTransitionTime":"2025-12-01T15:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.909871 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.922023 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.942611 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.961304 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.973867 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.987998 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.997317 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:26Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.998005 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.998113 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.998226 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.998325 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:26 crc kubenswrapper[4739]: I1201 15:35:26.998436 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:26Z","lastTransitionTime":"2025-12-01T15:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.010513 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.032537 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7201f18006d89a4fca89142fe792cc3def0fa026eae9d587af724f777a746805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.042898 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.052702 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.065850 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.078975 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.090492 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.101598 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.101650 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.101660 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.101675 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.101685 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:27Z","lastTransitionTime":"2025-12-01T15:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.102586 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.119761 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.148145 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.166144 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.177486 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.203999 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.204057 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.204070 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.204088 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.204102 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:27Z","lastTransitionTime":"2025-12-01T15:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.307109 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.307178 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.307194 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.307216 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.307232 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:27Z","lastTransitionTime":"2025-12-01T15:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.409827 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.409865 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.409901 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.409919 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.409932 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:27Z","lastTransitionTime":"2025-12-01T15:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.532987 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.533012 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.533059 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:27 crc kubenswrapper[4739]: E1201 15:35:27.533097 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:27 crc kubenswrapper[4739]: E1201 15:35:27.533247 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:27 crc kubenswrapper[4739]: E1201 15:35:27.533407 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.534247 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.534277 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.534293 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.534314 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.534331 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:27Z","lastTransitionTime":"2025-12-01T15:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.636268 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.636347 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.636368 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.636393 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.636410 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:27Z","lastTransitionTime":"2025-12-01T15:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.688885 4739 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.689555 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.714482 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.727871 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.738734 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.738789 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.738808 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.738831 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.738849 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:27Z","lastTransitionTime":"2025-12-01T15:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.749374 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.762059 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.779628 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.794551 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.815116 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.836196 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.841082 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.841132 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.841150 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.841173 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.841190 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:27Z","lastTransitionTime":"2025-12-01T15:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.855993 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.877564 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.900015 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.943525 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7201f18006d89a4fca89142fe792cc3def0fa026eae9d587af724f777a746805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.943863 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.943881 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.943890 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.943904 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.943912 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:27Z","lastTransitionTime":"2025-12-01T15:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.959653 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.980425 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:27 crc kubenswrapper[4739]: I1201 15:35:27.990925 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.000836 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:27Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.045643 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.045688 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.045697 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.045727 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.045737 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:28Z","lastTransitionTime":"2025-12-01T15:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.149121 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.149157 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.149168 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.149185 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.149195 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:28Z","lastTransitionTime":"2025-12-01T15:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.252093 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.252148 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.252166 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.252191 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.252208 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:28Z","lastTransitionTime":"2025-12-01T15:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.355084 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.355151 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.355166 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.355186 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.355197 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:28Z","lastTransitionTime":"2025-12-01T15:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.464128 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.464198 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.464216 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.464239 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.464255 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:28Z","lastTransitionTime":"2025-12-01T15:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.493535 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.504143 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.523000 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.538745 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.549255 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.566229 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.566269 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.566282 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.566300 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.566312 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:28Z","lastTransitionTime":"2025-12-01T15:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.569403 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.584096 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.603792 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.623907 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.643402 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.669580 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.669636 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.669647 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.669664 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.669676 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:28Z","lastTransitionTime":"2025-12-01T15:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.671268 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7201f18006d89a4fca89142fe792cc3def0fa026eae9d587af724f777a746805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.692343 4739 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.701790 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.720393 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.735339 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.753773 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.772982 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.773038 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.773055 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.773080 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.773098 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:28Z","lastTransitionTime":"2025-12-01T15:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.875839 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.875912 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.875939 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.875963 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.875980 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:28Z","lastTransitionTime":"2025-12-01T15:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.979857 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.979919 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.979940 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.979963 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:28 crc kubenswrapper[4739]: I1201 15:35:28.979982 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:28Z","lastTransitionTime":"2025-12-01T15:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.083229 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.083302 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.083326 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.083355 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.083378 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:29Z","lastTransitionTime":"2025-12-01T15:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.186116 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.186181 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.186198 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.186221 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.186237 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:29Z","lastTransitionTime":"2025-12-01T15:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.288651 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.288709 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.288727 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.288783 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.288806 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:29Z","lastTransitionTime":"2025-12-01T15:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.391272 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.391337 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.391362 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.391393 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.391452 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:29Z","lastTransitionTime":"2025-12-01T15:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.476346 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.476385 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:29 crc kubenswrapper[4739]: E1201 15:35:29.476555 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.476661 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:29 crc kubenswrapper[4739]: E1201 15:35:29.476888 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:29 crc kubenswrapper[4739]: E1201 15:35:29.477001 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.494565 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.494640 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.494666 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.494695 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.494719 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:29Z","lastTransitionTime":"2025-12-01T15:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.597585 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.597647 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.597664 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.597688 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.597705 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:29Z","lastTransitionTime":"2025-12-01T15:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.697858 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovnkube-controller/0.log" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.700397 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.700512 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.700555 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.700587 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.700609 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:29Z","lastTransitionTime":"2025-12-01T15:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.702935 4739 generic.go:334] "Generic (PLEG): container finished" podID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerID="7201f18006d89a4fca89142fe792cc3def0fa026eae9d587af724f777a746805" exitCode=1 Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.703049 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerDied","Data":"7201f18006d89a4fca89142fe792cc3def0fa026eae9d587af724f777a746805"} Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.704265 4739 scope.go:117] "RemoveContainer" containerID="7201f18006d89a4fca89142fe792cc3def0fa026eae9d587af724f777a746805" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.747229 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.769554 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.788037 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.841585 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.841623 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.841632 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.841646 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.841656 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:29Z","lastTransitionTime":"2025-12-01T15:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.845081 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.859177 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.872383 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.888101 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.902186 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.914588 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.926947 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.937306 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.944325 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.944390 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.944410 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.944453 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.944470 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:29Z","lastTransitionTime":"2025-12-01T15:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.949433 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.970207 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7201f18006d89a4fca89142fe792cc3def0fa026eae9d587af724f777a746805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7201f18006d89a4fca89142fe792cc3def0fa026eae9d587af724f777a746805\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:28Z\\\",\\\"message\\\":\\\"v1.Namespace event handler 1 for removal\\\\nI1201 15:35:28.473775 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 15:35:28.473797 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 15:35:28.473832 6034 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.473854 6034 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:28.473872 6034 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:28.473887 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:28.473910 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 15:35:28.474160 6034 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.473836 6034 factory.go:656] Stopping watch factory\\\\nI1201 15:35:28.474462 6034 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.475102 6034 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.475315 6034 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:29 crc kubenswrapper[4739]: I1201 15:35:29.997775 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:29Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.020374 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:30Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.047103 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.047136 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.047144 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.047158 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.047169 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:30Z","lastTransitionTime":"2025-12-01T15:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.149554 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.149608 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.149626 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.149648 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.149661 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:30Z","lastTransitionTime":"2025-12-01T15:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.251607 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.251643 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.251654 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.251669 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.251679 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:30Z","lastTransitionTime":"2025-12-01T15:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.354042 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.354069 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.354078 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.354093 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.354101 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:30Z","lastTransitionTime":"2025-12-01T15:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.456526 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.456595 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.456609 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.456633 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.456647 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:30Z","lastTransitionTime":"2025-12-01T15:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.559499 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.559559 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.559578 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.559608 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.559626 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:30Z","lastTransitionTime":"2025-12-01T15:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.662668 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.662744 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.662757 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.662781 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.662798 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:30Z","lastTransitionTime":"2025-12-01T15:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.715039 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovnkube-controller/0.log" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.719979 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerStarted","Data":"796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f"} Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.720170 4739 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.740765 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:30Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.759111 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:30Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.765301 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.765378 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.765398 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.765452 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.765471 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:30Z","lastTransitionTime":"2025-12-01T15:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.776715 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:30Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.791484 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.792237 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:30Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.818216 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7201f18006d89a4fca89142fe792cc3def0fa026eae9d587af724f777a746805\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:28Z\\\",\\\"message\\\":\\\"v1.Namespace event handler 1 for removal\\\\nI1201 15:35:28.473775 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 15:35:28.473797 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 15:35:28.473832 6034 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.473854 6034 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:28.473872 6034 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:28.473887 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:28.473910 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 15:35:28.474160 6034 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.473836 6034 factory.go:656] Stopping watch factory\\\\nI1201 15:35:28.474462 6034 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.475102 6034 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.475315 6034 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:30Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.837039 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:30Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.851853 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:30Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.867580 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.867640 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.867651 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.867667 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.867678 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:30Z","lastTransitionTime":"2025-12-01T15:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.874404 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:30Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.892341 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:30Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.905839 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:30Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.929851 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:30Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.945847 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:30Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.966064 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:30Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.969887 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.969926 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.969935 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.969950 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.969960 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:30Z","lastTransitionTime":"2025-12-01T15:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:30 crc kubenswrapper[4739]: I1201 15:35:30.981240 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:30Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.003942 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.071848 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.072102 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.072220 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.072346 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.072493 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:31Z","lastTransitionTime":"2025-12-01T15:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.175795 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.175846 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.175864 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.175886 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.175902 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:31Z","lastTransitionTime":"2025-12-01T15:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.268608 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.279675 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.279728 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.279741 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.279764 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.279780 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:31Z","lastTransitionTime":"2025-12-01T15:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.307936 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.324058 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.339940 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.354101 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.370248 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.382350 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.382397 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.382410 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.382444 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.382459 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:31Z","lastTransitionTime":"2025-12-01T15:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.387277 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.398929 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.410728 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.429881 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.452520 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7201f18006d89a4fca89142fe792cc3def0fa026eae9d587af724f777a746805\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:28Z\\\",\\\"message\\\":\\\"v1.Namespace event handler 1 for removal\\\\nI1201 15:35:28.473775 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 15:35:28.473797 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 15:35:28.473832 6034 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.473854 6034 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:28.473872 6034 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:28.473887 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:28.473910 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 15:35:28.474160 6034 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.473836 6034 factory.go:656] Stopping watch factory\\\\nI1201 15:35:28.474462 6034 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.475102 6034 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.475315 6034 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.466312 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.476994 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.477014 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.476994 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:31 crc kubenswrapper[4739]: E1201 15:35:31.477120 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:31 crc kubenswrapper[4739]: E1201 15:35:31.477184 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:31 crc kubenswrapper[4739]: E1201 15:35:31.477292 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.477675 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.485261 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.485310 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.485321 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.485337 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.485350 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:31Z","lastTransitionTime":"2025-12-01T15:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.495676 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.509963 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.524787 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.552048 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.552097 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.552112 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.552132 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.552146 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:31Z","lastTransitionTime":"2025-12-01T15:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:31 crc kubenswrapper[4739]: E1201 15:35:31.571979 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.576792 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.576851 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.576878 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.576908 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.576931 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:31Z","lastTransitionTime":"2025-12-01T15:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:31 crc kubenswrapper[4739]: E1201 15:35:31.597784 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.602765 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.602814 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.602828 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.602852 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.602868 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:31Z","lastTransitionTime":"2025-12-01T15:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:31 crc kubenswrapper[4739]: E1201 15:35:31.624895 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.629880 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.629990 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.630008 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.630036 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.630059 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:31Z","lastTransitionTime":"2025-12-01T15:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:31 crc kubenswrapper[4739]: E1201 15:35:31.652225 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.652975 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c"] Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.653875 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.657047 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.658380 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.658474 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.658497 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.658523 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.658530 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.658541 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:31Z","lastTransitionTime":"2025-12-01T15:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.670639 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw9z6\" (UniqueName: \"kubernetes.io/projected/32ef9ece-a823-470c-8c9c-a3df83efa972-kube-api-access-xw9z6\") pod \"ovnkube-control-plane-749d76644c-ftp8c\" (UID: \"32ef9ece-a823-470c-8c9c-a3df83efa972\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.670842 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/32ef9ece-a823-470c-8c9c-a3df83efa972-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ftp8c\" (UID: \"32ef9ece-a823-470c-8c9c-a3df83efa972\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.670907 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/32ef9ece-a823-470c-8c9c-a3df83efa972-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ftp8c\" (UID: \"32ef9ece-a823-470c-8c9c-a3df83efa972\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.671004 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/32ef9ece-a823-470c-8c9c-a3df83efa972-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ftp8c\" (UID: \"32ef9ece-a823-470c-8c9c-a3df83efa972\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" Dec 01 15:35:31 crc kubenswrapper[4739]: E1201 15:35:31.684688 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: E1201 15:35:31.685037 4739 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.687588 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.687654 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.687674 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.687697 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.687715 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:31Z","lastTransitionTime":"2025-12-01T15:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.691727 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.714351 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.726460 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovnkube-controller/1.log" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.727520 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovnkube-controller/0.log" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.731330 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.733084 4739 generic.go:334] "Generic (PLEG): container finished" podID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerID="796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f" exitCode=1 Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.733156 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerDied","Data":"796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f"} Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.733231 4739 scope.go:117] "RemoveContainer" containerID="7201f18006d89a4fca89142fe792cc3def0fa026eae9d587af724f777a746805" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.734603 4739 scope.go:117] "RemoveContainer" containerID="796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f" Dec 01 15:35:31 crc kubenswrapper[4739]: E1201 15:35:31.734899 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.752243 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.769271 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.772145 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/32ef9ece-a823-470c-8c9c-a3df83efa972-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ftp8c\" (UID: \"32ef9ece-a823-470c-8c9c-a3df83efa972\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.772245 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/32ef9ece-a823-470c-8c9c-a3df83efa972-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ftp8c\" (UID: \"32ef9ece-a823-470c-8c9c-a3df83efa972\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.772319 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/32ef9ece-a823-470c-8c9c-a3df83efa972-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ftp8c\" (UID: \"32ef9ece-a823-470c-8c9c-a3df83efa972\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.772401 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw9z6\" (UniqueName: \"kubernetes.io/projected/32ef9ece-a823-470c-8c9c-a3df83efa972-kube-api-access-xw9z6\") pod \"ovnkube-control-plane-749d76644c-ftp8c\" (UID: \"32ef9ece-a823-470c-8c9c-a3df83efa972\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.773122 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/32ef9ece-a823-470c-8c9c-a3df83efa972-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ftp8c\" (UID: \"32ef9ece-a823-470c-8c9c-a3df83efa972\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.773208 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/32ef9ece-a823-470c-8c9c-a3df83efa972-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ftp8c\" (UID: \"32ef9ece-a823-470c-8c9c-a3df83efa972\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.781377 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/32ef9ece-a823-470c-8c9c-a3df83efa972-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ftp8c\" (UID: \"32ef9ece-a823-470c-8c9c-a3df83efa972\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.790323 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.790371 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.790389 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.790447 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.790467 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:31Z","lastTransitionTime":"2025-12-01T15:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.795903 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw9z6\" (UniqueName: \"kubernetes.io/projected/32ef9ece-a823-470c-8c9c-a3df83efa972-kube-api-access-xw9z6\") pod \"ovnkube-control-plane-749d76644c-ftp8c\" (UID: \"32ef9ece-a823-470c-8c9c-a3df83efa972\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.800140 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7201f18006d89a4fca89142fe792cc3def0fa026eae9d587af724f777a746805\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:28Z\\\",\\\"message\\\":\\\"v1.Namespace event handler 1 for removal\\\\nI1201 15:35:28.473775 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 15:35:28.473797 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 15:35:28.473832 6034 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.473854 6034 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:28.473872 6034 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:28.473887 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:28.473910 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 15:35:28.474160 6034 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.473836 6034 factory.go:656] Stopping watch factory\\\\nI1201 15:35:28.474462 6034 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.475102 6034 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.475315 6034 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.817410 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.830892 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.849666 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.869936 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.887265 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.893930 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.894001 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.894026 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.894060 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.894085 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:31Z","lastTransitionTime":"2025-12-01T15:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.911199 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.926679 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.943919 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.960067 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.975780 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.976501 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.994726 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:31Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.996386 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.996453 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.996470 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.996490 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:31 crc kubenswrapper[4739]: I1201 15:35:31.996506 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:31Z","lastTransitionTime":"2025-12-01T15:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.011925 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.040767 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.056356 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.092815 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7201f18006d89a4fca89142fe792cc3def0fa026eae9d587af724f777a746805\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:28Z\\\",\\\"message\\\":\\\"v1.Namespace event handler 1 for removal\\\\nI1201 15:35:28.473775 6034 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 15:35:28.473797 6034 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 15:35:28.473832 6034 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.473854 6034 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:28.473872 6034 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:28.473887 6034 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:28.473910 6034 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 15:35:28.474160 6034 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.473836 6034 factory.go:656] Stopping watch factory\\\\nI1201 15:35:28.474462 6034 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.475102 6034 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 15:35:28.475315 6034 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"val\\\\nI1201 15:35:30.823819 6160 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:30.823835 6160 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 15:35:30.823846 6160 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 15:35:30.823851 6160 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 15:35:30.823862 6160 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:30.823867 6160 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:30.823878 6160 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 15:35:30.823895 6160 factory.go:656] Stopping watch factory\\\\nI1201 15:35:30.823900 6160 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:30.823895 6160 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:30.823911 6160 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:30.823918 6160 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:30.823926 6160 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:30.823937 6160 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:30.823953 6160 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:30.823961 6160 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.099385 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.099477 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.099504 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.099537 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.099564 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:32Z","lastTransitionTime":"2025-12-01T15:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.108905 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.125297 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.147900 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.164643 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.179021 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.195312 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.202672 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.202721 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.202739 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.202762 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.202779 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:32Z","lastTransitionTime":"2025-12-01T15:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.220313 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.240783 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.259572 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.285151 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.304893 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.305739 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.305820 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.305846 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.305878 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.305906 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:32Z","lastTransitionTime":"2025-12-01T15:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.409545 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.409601 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.409641 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.409659 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.409671 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:32Z","lastTransitionTime":"2025-12-01T15:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.512888 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.512932 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.512944 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.512963 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.512975 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:32Z","lastTransitionTime":"2025-12-01T15:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.577498 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.577617 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:32 crc kubenswrapper[4739]: E1201 15:35:32.577710 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:35:48.577692421 +0000 UTC m=+50.403438515 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:35:32 crc kubenswrapper[4739]: E1201 15:35:32.577800 4739 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:35:32 crc kubenswrapper[4739]: E1201 15:35:32.577880 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:48.577863687 +0000 UTC m=+50.403609791 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.615315 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.615358 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.615371 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.615388 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.615401 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:32Z","lastTransitionTime":"2025-12-01T15:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.678741 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.678790 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.678823 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:32 crc kubenswrapper[4739]: E1201 15:35:32.678887 4739 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:35:32 crc kubenswrapper[4739]: E1201 15:35:32.678915 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:35:32 crc kubenswrapper[4739]: E1201 15:35:32.678928 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:35:32 crc kubenswrapper[4739]: E1201 15:35:32.678938 4739 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:32 crc kubenswrapper[4739]: E1201 15:35:32.678951 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:48.678933749 +0000 UTC m=+50.504679863 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:35:32 crc kubenswrapper[4739]: E1201 15:35:32.678969 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:48.67896107 +0000 UTC m=+50.504707184 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:32 crc kubenswrapper[4739]: E1201 15:35:32.678989 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:35:32 crc kubenswrapper[4739]: E1201 15:35:32.678998 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:35:32 crc kubenswrapper[4739]: E1201 15:35:32.679005 4739 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:32 crc kubenswrapper[4739]: E1201 15:35:32.679025 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 15:35:48.679016892 +0000 UTC m=+50.504762986 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.718570 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.718614 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.718625 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.718640 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.718651 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:32Z","lastTransitionTime":"2025-12-01T15:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.737920 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovnkube-controller/1.log" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.742033 4739 scope.go:117] "RemoveContainer" containerID="796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f" Dec 01 15:35:32 crc kubenswrapper[4739]: E1201 15:35:32.742165 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.743285 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" event={"ID":"32ef9ece-a823-470c-8c9c-a3df83efa972","Type":"ContainerStarted","Data":"bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c"} Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.743352 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" event={"ID":"32ef9ece-a823-470c-8c9c-a3df83efa972","Type":"ContainerStarted","Data":"071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44"} Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.743365 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" event={"ID":"32ef9ece-a823-470c-8c9c-a3df83efa972","Type":"ContainerStarted","Data":"ec1fecef243d79fdb3f540b3f42d6ab34a2d9a5fb12f0d631c678da69a2e020d"} Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.766771 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.790109 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.813343 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.822183 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.822219 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.822228 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.822243 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.822254 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:32Z","lastTransitionTime":"2025-12-01T15:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.835083 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.857250 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.888514 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"val\\\\nI1201 15:35:30.823819 6160 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:30.823835 6160 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 15:35:30.823846 6160 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 15:35:30.823851 6160 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 15:35:30.823862 6160 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:30.823867 6160 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:30.823878 6160 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 15:35:30.823895 6160 factory.go:656] Stopping watch factory\\\\nI1201 15:35:30.823900 6160 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:30.823895 6160 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:30.823911 6160 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:30.823918 6160 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:30.823926 6160 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:30.823937 6160 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:30.823953 6160 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:30.823961 6160 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.907473 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.926212 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.926269 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.926287 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.926313 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.926332 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:32Z","lastTransitionTime":"2025-12-01T15:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.927976 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.961046 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:32 crc kubenswrapper[4739]: I1201 15:35:32.983576 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:32Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.002658 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.025235 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.029747 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.029823 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.029844 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.029871 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.029891 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:33Z","lastTransitionTime":"2025-12-01T15:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.044971 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.067808 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.087263 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.110912 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.128106 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.132344 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.132408 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.132440 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.132467 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.132481 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:33Z","lastTransitionTime":"2025-12-01T15:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.144778 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.164015 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.178833 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.180367 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-bsjx4"] Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.183658 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:33 crc kubenswrapper[4739]: E1201 15:35:33.183815 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.186201 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs\") pod \"network-metrics-daemon-bsjx4\" (UID: \"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\") " pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.186253 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts24m\" (UniqueName: \"kubernetes.io/projected/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-kube-api-access-ts24m\") pod \"network-metrics-daemon-bsjx4\" (UID: \"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\") " pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.195789 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.212772 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.227381 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.236335 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.236389 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.236407 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.236461 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.236479 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:33Z","lastTransitionTime":"2025-12-01T15:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.241751 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.256732 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.276621 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.287042 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs\") pod \"network-metrics-daemon-bsjx4\" (UID: \"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\") " pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.287332 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts24m\" (UniqueName: \"kubernetes.io/projected/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-kube-api-access-ts24m\") pod \"network-metrics-daemon-bsjx4\" (UID: \"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\") " pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:33 crc kubenswrapper[4739]: E1201 15:35:33.287743 4739 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:35:33 crc kubenswrapper[4739]: E1201 15:35:33.287853 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs podName:55ed5d4d-1f46-4e38-9176-2d1a8e5417cf nodeName:}" failed. No retries permitted until 2025-12-01 15:35:33.787822809 +0000 UTC m=+35.613568933 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs") pod "network-metrics-daemon-bsjx4" (UID: "55ed5d4d-1f46-4e38-9176-2d1a8e5417cf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.299392 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"val\\\\nI1201 15:35:30.823819 6160 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:30.823835 6160 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 15:35:30.823846 6160 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 15:35:30.823851 6160 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 15:35:30.823862 6160 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:30.823867 6160 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:30.823878 6160 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 15:35:30.823895 6160 factory.go:656] Stopping watch factory\\\\nI1201 15:35:30.823900 6160 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:30.823895 6160 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:30.823911 6160 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:30.823918 6160 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:30.823926 6160 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:30.823937 6160 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:30.823953 6160 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:30.823961 6160 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.320299 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.325144 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts24m\" (UniqueName: \"kubernetes.io/projected/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-kube-api-access-ts24m\") pod \"network-metrics-daemon-bsjx4\" (UID: \"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\") " pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.336115 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.340018 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.340099 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.340118 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.340631 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.340697 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:33Z","lastTransitionTime":"2025-12-01T15:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.351238 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.366182 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.381308 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.396078 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.415854 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.432954 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.444668 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.444733 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.444751 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.444780 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.444798 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:33Z","lastTransitionTime":"2025-12-01T15:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.460339 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.476639 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.476748 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:33 crc kubenswrapper[4739]: E1201 15:35:33.476811 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:33 crc kubenswrapper[4739]: E1201 15:35:33.476932 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.477052 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:33 crc kubenswrapper[4739]: E1201 15:35:33.477145 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.483145 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.497066 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.511477 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.526816 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.547967 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.548011 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.548027 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.548048 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.548062 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:33Z","lastTransitionTime":"2025-12-01T15:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.557652 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"val\\\\nI1201 15:35:30.823819 6160 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:30.823835 6160 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 15:35:30.823846 6160 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 15:35:30.823851 6160 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 15:35:30.823862 6160 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:30.823867 6160 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:30.823878 6160 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 15:35:30.823895 6160 factory.go:656] Stopping watch factory\\\\nI1201 15:35:30.823900 6160 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:30.823895 6160 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:30.823911 6160 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:30.823918 6160 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:30.823926 6160 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:30.823937 6160 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:30.823953 6160 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:30.823961 6160 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.574785 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.587518 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.605211 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.625972 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.645346 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.650785 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.650852 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.650870 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.650894 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.650915 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:33Z","lastTransitionTime":"2025-12-01T15:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.680444 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.701901 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.721132 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:33Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.754381 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.754467 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.754483 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.754503 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.754520 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:33Z","lastTransitionTime":"2025-12-01T15:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.792563 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs\") pod \"network-metrics-daemon-bsjx4\" (UID: \"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\") " pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:33 crc kubenswrapper[4739]: E1201 15:35:33.792761 4739 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:35:33 crc kubenswrapper[4739]: E1201 15:35:33.792851 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs podName:55ed5d4d-1f46-4e38-9176-2d1a8e5417cf nodeName:}" failed. No retries permitted until 2025-12-01 15:35:34.792828687 +0000 UTC m=+36.618574821 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs") pod "network-metrics-daemon-bsjx4" (UID: "55ed5d4d-1f46-4e38-9176-2d1a8e5417cf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.856966 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.857162 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.857246 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.857319 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.857375 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:33Z","lastTransitionTime":"2025-12-01T15:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.959728 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.959805 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.959824 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.959847 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:33 crc kubenswrapper[4739]: I1201 15:35:33.959863 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:33Z","lastTransitionTime":"2025-12-01T15:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.062812 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.062840 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.062849 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.062863 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.062872 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:34Z","lastTransitionTime":"2025-12-01T15:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.165677 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.165743 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.165757 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.165774 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.165785 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:34Z","lastTransitionTime":"2025-12-01T15:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.269163 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.269226 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.269243 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.269267 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.269284 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:34Z","lastTransitionTime":"2025-12-01T15:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.372745 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.372816 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.372838 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.372867 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.372889 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:34Z","lastTransitionTime":"2025-12-01T15:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.475961 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.476000 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.476009 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.476023 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.476033 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:34Z","lastTransitionTime":"2025-12-01T15:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.578155 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.578226 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.578245 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.578269 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.578286 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:34Z","lastTransitionTime":"2025-12-01T15:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.681103 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.681182 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.681208 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.681239 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.681261 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:34Z","lastTransitionTime":"2025-12-01T15:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.785050 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.785108 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.785126 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.785149 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.785166 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:34Z","lastTransitionTime":"2025-12-01T15:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.803248 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs\") pod \"network-metrics-daemon-bsjx4\" (UID: \"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\") " pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:34 crc kubenswrapper[4739]: E1201 15:35:34.803468 4739 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:35:34 crc kubenswrapper[4739]: E1201 15:35:34.803581 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs podName:55ed5d4d-1f46-4e38-9176-2d1a8e5417cf nodeName:}" failed. No retries permitted until 2025-12-01 15:35:36.803553006 +0000 UTC m=+38.629299140 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs") pod "network-metrics-daemon-bsjx4" (UID: "55ed5d4d-1f46-4e38-9176-2d1a8e5417cf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.888062 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.888122 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.888140 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.888168 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.888187 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:34Z","lastTransitionTime":"2025-12-01T15:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.991539 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.991583 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.991595 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.991612 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:34 crc kubenswrapper[4739]: I1201 15:35:34.991623 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:34Z","lastTransitionTime":"2025-12-01T15:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.094787 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.094869 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.094893 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.094921 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.094938 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:35Z","lastTransitionTime":"2025-12-01T15:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.197761 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.197820 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.197840 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.197863 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.197882 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:35Z","lastTransitionTime":"2025-12-01T15:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.301127 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.301182 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.301205 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.301232 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.301253 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:35Z","lastTransitionTime":"2025-12-01T15:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.403500 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.403558 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.403574 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.403593 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.403607 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:35Z","lastTransitionTime":"2025-12-01T15:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.477091 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.477155 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.477161 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:35 crc kubenswrapper[4739]: E1201 15:35:35.477248 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.477092 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:35 crc kubenswrapper[4739]: E1201 15:35:35.477533 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:35 crc kubenswrapper[4739]: E1201 15:35:35.477607 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:35 crc kubenswrapper[4739]: E1201 15:35:35.477693 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.506974 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.507040 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.507071 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.507115 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.507137 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:35Z","lastTransitionTime":"2025-12-01T15:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.610887 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.610936 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.610947 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.610967 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.610982 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:35Z","lastTransitionTime":"2025-12-01T15:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.713595 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.713634 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.713643 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.713656 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.713666 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:35Z","lastTransitionTime":"2025-12-01T15:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.815357 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.815467 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.815496 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.815525 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.815549 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:35Z","lastTransitionTime":"2025-12-01T15:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.919227 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.919293 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.919310 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.919334 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:35 crc kubenswrapper[4739]: I1201 15:35:35.919352 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:35Z","lastTransitionTime":"2025-12-01T15:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.022289 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.022350 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.022367 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.022464 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.022484 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:36Z","lastTransitionTime":"2025-12-01T15:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.124910 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.124971 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.124989 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.125015 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.125032 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:36Z","lastTransitionTime":"2025-12-01T15:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.227960 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.228033 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.228045 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.228065 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.228078 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:36Z","lastTransitionTime":"2025-12-01T15:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.331152 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.331216 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.331250 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.331281 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.331303 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:36Z","lastTransitionTime":"2025-12-01T15:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.439077 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.439134 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.439152 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.439176 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.439192 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:36Z","lastTransitionTime":"2025-12-01T15:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.542257 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.542322 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.542343 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.542372 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.542391 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:36Z","lastTransitionTime":"2025-12-01T15:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.645513 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.645853 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.646061 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.646271 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.646522 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:36Z","lastTransitionTime":"2025-12-01T15:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.749336 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.749643 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.749692 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.749721 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.749739 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:36Z","lastTransitionTime":"2025-12-01T15:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.826044 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs\") pod \"network-metrics-daemon-bsjx4\" (UID: \"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\") " pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:36 crc kubenswrapper[4739]: E1201 15:35:36.826339 4739 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:35:36 crc kubenswrapper[4739]: E1201 15:35:36.826518 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs podName:55ed5d4d-1f46-4e38-9176-2d1a8e5417cf nodeName:}" failed. No retries permitted until 2025-12-01 15:35:40.826476642 +0000 UTC m=+42.652222786 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs") pod "network-metrics-daemon-bsjx4" (UID: "55ed5d4d-1f46-4e38-9176-2d1a8e5417cf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.853052 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.853122 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.853146 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.853177 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.853201 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:36Z","lastTransitionTime":"2025-12-01T15:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.956563 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.956624 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.956647 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.956677 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:36 crc kubenswrapper[4739]: I1201 15:35:36.956697 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:36Z","lastTransitionTime":"2025-12-01T15:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.060080 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.060114 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.060123 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.060138 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.060147 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:37Z","lastTransitionTime":"2025-12-01T15:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.163014 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.163058 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.163069 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.163089 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.163133 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:37Z","lastTransitionTime":"2025-12-01T15:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.265659 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.265692 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.265703 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.265715 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.265723 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:37Z","lastTransitionTime":"2025-12-01T15:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.368577 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.368616 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.368625 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.368641 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.368651 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:37Z","lastTransitionTime":"2025-12-01T15:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.471178 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.471267 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.471294 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.471327 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.471354 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:37Z","lastTransitionTime":"2025-12-01T15:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.476888 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.476926 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:37 crc kubenswrapper[4739]: E1201 15:35:37.476992 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.476886 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.476895 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:37 crc kubenswrapper[4739]: E1201 15:35:37.477088 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:37 crc kubenswrapper[4739]: E1201 15:35:37.477212 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:37 crc kubenswrapper[4739]: E1201 15:35:37.477287 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.573899 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.573939 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.573951 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.573970 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.573982 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:37Z","lastTransitionTime":"2025-12-01T15:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.676541 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.676616 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.676628 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.676643 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.676655 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:37Z","lastTransitionTime":"2025-12-01T15:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.779026 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.779088 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.779111 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.779140 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.779162 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:37Z","lastTransitionTime":"2025-12-01T15:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.881628 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.881663 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.881673 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.881691 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.881706 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:37Z","lastTransitionTime":"2025-12-01T15:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.983622 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.983663 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.983675 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.983690 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:37 crc kubenswrapper[4739]: I1201 15:35:37.983704 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:37Z","lastTransitionTime":"2025-12-01T15:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.086801 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.086863 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.086879 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.086943 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.086959 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:38Z","lastTransitionTime":"2025-12-01T15:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.189887 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.189936 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.189948 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.189968 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.189980 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:38Z","lastTransitionTime":"2025-12-01T15:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.293184 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.293254 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.293267 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.293288 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.293300 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:38Z","lastTransitionTime":"2025-12-01T15:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.395545 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.395582 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.395594 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.395610 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.395623 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:38Z","lastTransitionTime":"2025-12-01T15:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.497798 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.498288 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.498323 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.498333 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.498347 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.498365 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:38Z","lastTransitionTime":"2025-12-01T15:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.511945 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.524105 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.541902 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.564015 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.586762 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"val\\\\nI1201 15:35:30.823819 6160 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:30.823835 6160 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 15:35:30.823846 6160 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 15:35:30.823851 6160 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 15:35:30.823862 6160 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:30.823867 6160 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:30.823878 6160 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 15:35:30.823895 6160 factory.go:656] Stopping watch factory\\\\nI1201 15:35:30.823900 6160 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:30.823895 6160 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:30.823911 6160 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:30.823918 6160 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:30.823926 6160 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:30.823937 6160 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:30.823953 6160 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:30.823961 6160 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.602005 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.602054 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.602064 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.602078 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.602090 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:38Z","lastTransitionTime":"2025-12-01T15:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.603081 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.616718 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.634743 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.650773 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.667772 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.687508 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.702662 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.703798 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.703825 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.703835 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.703853 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.703864 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:38Z","lastTransitionTime":"2025-12-01T15:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.719884 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.733643 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.749996 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.764741 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:38Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.806098 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.806183 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.806204 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.806227 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.806243 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:38Z","lastTransitionTime":"2025-12-01T15:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.908849 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.908907 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.908926 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.908949 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:38 crc kubenswrapper[4739]: I1201 15:35:38.908968 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:38Z","lastTransitionTime":"2025-12-01T15:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.034232 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.034575 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.034746 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.034878 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.035009 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:39Z","lastTransitionTime":"2025-12-01T15:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.138488 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.138769 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.138936 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.139063 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.139177 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:39Z","lastTransitionTime":"2025-12-01T15:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.247672 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.247755 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.247782 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.247811 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.247832 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:39Z","lastTransitionTime":"2025-12-01T15:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.350320 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.350373 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.350390 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.350406 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.350437 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:39Z","lastTransitionTime":"2025-12-01T15:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.453283 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.453358 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.453380 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.453410 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.453465 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:39Z","lastTransitionTime":"2025-12-01T15:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.477034 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.477069 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.477085 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:39 crc kubenswrapper[4739]: E1201 15:35:39.477569 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.477156 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:39 crc kubenswrapper[4739]: E1201 15:35:39.477713 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:39 crc kubenswrapper[4739]: E1201 15:35:39.477969 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:35:39 crc kubenswrapper[4739]: E1201 15:35:39.478159 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.556694 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.556752 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.556767 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.556789 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.556804 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:39Z","lastTransitionTime":"2025-12-01T15:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.664205 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.664259 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.664277 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.664300 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.664316 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:39Z","lastTransitionTime":"2025-12-01T15:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.768903 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.768966 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.768985 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.769013 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.769034 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:39Z","lastTransitionTime":"2025-12-01T15:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.872853 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.872910 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.872930 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.872956 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.872975 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:39Z","lastTransitionTime":"2025-12-01T15:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.975350 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.975445 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.975462 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.975488 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:39 crc kubenswrapper[4739]: I1201 15:35:39.975504 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:39Z","lastTransitionTime":"2025-12-01T15:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.078952 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.079044 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.079070 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.079096 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.079112 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:40Z","lastTransitionTime":"2025-12-01T15:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.182647 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.182710 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.182728 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.182753 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.182773 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:40Z","lastTransitionTime":"2025-12-01T15:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.286249 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.286314 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.286332 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.286357 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.286375 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:40Z","lastTransitionTime":"2025-12-01T15:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.388792 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.388868 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.388889 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.388906 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.388915 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:40Z","lastTransitionTime":"2025-12-01T15:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.490742 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.490805 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.490827 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.490857 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.490879 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:40Z","lastTransitionTime":"2025-12-01T15:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.593372 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.593473 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.593498 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.593520 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.593534 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:40Z","lastTransitionTime":"2025-12-01T15:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.696660 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.696706 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.696737 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.696757 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.696768 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:40Z","lastTransitionTime":"2025-12-01T15:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.799758 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.799817 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.799835 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.799860 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.799877 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:40Z","lastTransitionTime":"2025-12-01T15:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.871046 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs\") pod \"network-metrics-daemon-bsjx4\" (UID: \"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\") " pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:40 crc kubenswrapper[4739]: E1201 15:35:40.871290 4739 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:35:40 crc kubenswrapper[4739]: E1201 15:35:40.871358 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs podName:55ed5d4d-1f46-4e38-9176-2d1a8e5417cf nodeName:}" failed. No retries permitted until 2025-12-01 15:35:48.871337312 +0000 UTC m=+50.697083436 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs") pod "network-metrics-daemon-bsjx4" (UID: "55ed5d4d-1f46-4e38-9176-2d1a8e5417cf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.903245 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.903325 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.903343 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.903369 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:40 crc kubenswrapper[4739]: I1201 15:35:40.903388 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:40Z","lastTransitionTime":"2025-12-01T15:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.007304 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.007359 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.007380 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.007408 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.007464 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:41Z","lastTransitionTime":"2025-12-01T15:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.110440 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.110507 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.110526 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.110557 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.110581 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:41Z","lastTransitionTime":"2025-12-01T15:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.214319 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.214379 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.214398 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.214474 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.214504 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:41Z","lastTransitionTime":"2025-12-01T15:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.318618 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.318693 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.318720 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.318752 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.318779 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:41Z","lastTransitionTime":"2025-12-01T15:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.422368 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.422453 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.422511 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.422540 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.422557 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:41Z","lastTransitionTime":"2025-12-01T15:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.476844 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:41 crc kubenswrapper[4739]: E1201 15:35:41.478196 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.477889 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:41 crc kubenswrapper[4739]: E1201 15:35:41.478811 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.476844 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:41 crc kubenswrapper[4739]: E1201 15:35:41.479326 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.479638 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:41 crc kubenswrapper[4739]: E1201 15:35:41.479884 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.525964 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.526036 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.526056 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.526086 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.526110 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:41Z","lastTransitionTime":"2025-12-01T15:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.629757 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.629824 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.629841 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.629874 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.629896 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:41Z","lastTransitionTime":"2025-12-01T15:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.732977 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.733078 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.733098 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.733122 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.733138 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:41Z","lastTransitionTime":"2025-12-01T15:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.836094 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.836143 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.836160 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.836182 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.836200 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:41Z","lastTransitionTime":"2025-12-01T15:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.938862 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.938922 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.938939 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.938974 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.938991 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:41Z","lastTransitionTime":"2025-12-01T15:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.951407 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.951534 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.951564 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.951590 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.951610 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:41Z","lastTransitionTime":"2025-12-01T15:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:41 crc kubenswrapper[4739]: E1201 15:35:41.973896 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:41Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.979273 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.979355 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.979370 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.979393 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:41 crc kubenswrapper[4739]: I1201 15:35:41.979411 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:41Z","lastTransitionTime":"2025-12-01T15:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:42 crc kubenswrapper[4739]: E1201 15:35:42.000607 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:41Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.006391 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.006503 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.006522 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.006547 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.006567 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:42Z","lastTransitionTime":"2025-12-01T15:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:42 crc kubenswrapper[4739]: E1201 15:35:42.027291 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:42Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.033665 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.033726 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.033749 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.033782 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.033799 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:42Z","lastTransitionTime":"2025-12-01T15:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:42 crc kubenswrapper[4739]: E1201 15:35:42.054880 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:42Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.060597 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.060651 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.060673 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.060697 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.060717 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:42Z","lastTransitionTime":"2025-12-01T15:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:42 crc kubenswrapper[4739]: E1201 15:35:42.082773 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:42Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:42 crc kubenswrapper[4739]: E1201 15:35:42.083003 4739 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.085062 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.085100 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.085117 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.085140 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.085157 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:42Z","lastTransitionTime":"2025-12-01T15:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.188377 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.188537 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.188565 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.188598 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.188625 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:42Z","lastTransitionTime":"2025-12-01T15:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.291222 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.291287 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.291309 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.291339 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.291362 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:42Z","lastTransitionTime":"2025-12-01T15:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.394185 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.394269 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.394289 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.394317 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.394336 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:42Z","lastTransitionTime":"2025-12-01T15:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.497036 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.497100 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.497111 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.497126 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.497137 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:42Z","lastTransitionTime":"2025-12-01T15:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.600672 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.600729 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.600748 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.600772 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.600793 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:42Z","lastTransitionTime":"2025-12-01T15:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.704086 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.704145 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.704165 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.704221 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.704239 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:42Z","lastTransitionTime":"2025-12-01T15:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.807058 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.807136 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.807160 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.807188 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.807207 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:42Z","lastTransitionTime":"2025-12-01T15:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.910085 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.910144 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.910162 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.910187 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:42 crc kubenswrapper[4739]: I1201 15:35:42.910204 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:42Z","lastTransitionTime":"2025-12-01T15:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.014461 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.014567 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.014633 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.014662 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.014683 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:43Z","lastTransitionTime":"2025-12-01T15:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.118604 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.118656 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.118670 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.118689 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.118703 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:43Z","lastTransitionTime":"2025-12-01T15:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.221290 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.221347 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.221363 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.221386 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.221404 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:43Z","lastTransitionTime":"2025-12-01T15:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.325390 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.325505 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.325577 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.325647 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.325670 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:43Z","lastTransitionTime":"2025-12-01T15:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.430218 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.430318 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.430340 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.430374 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.430393 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:43Z","lastTransitionTime":"2025-12-01T15:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.476884 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.476919 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.477039 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:43 crc kubenswrapper[4739]: E1201 15:35:43.477222 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.477239 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:43 crc kubenswrapper[4739]: E1201 15:35:43.477661 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:43 crc kubenswrapper[4739]: E1201 15:35:43.477914 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:43 crc kubenswrapper[4739]: E1201 15:35:43.477878 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.478201 4739 scope.go:117] "RemoveContainer" containerID="796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.532785 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.533152 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.533171 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.533199 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.533220 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:43Z","lastTransitionTime":"2025-12-01T15:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.636101 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.636152 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.636169 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.636195 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.636212 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:43Z","lastTransitionTime":"2025-12-01T15:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.740100 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.740135 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.740146 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.740162 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.740174 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:43Z","lastTransitionTime":"2025-12-01T15:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.785689 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovnkube-controller/1.log" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.789950 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerStarted","Data":"a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667"} Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.790717 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.808816 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:43Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.839510 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:43Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.842344 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.842400 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.842416 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.842458 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.842472 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:43Z","lastTransitionTime":"2025-12-01T15:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.855478 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:43Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.868502 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:43Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.885729 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:43Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.899652 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:43Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.910674 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:43Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.926326 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:43Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.941671 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:43Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.945222 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.945264 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.945275 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.945293 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.945310 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:43Z","lastTransitionTime":"2025-12-01T15:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.959172 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:43Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.973666 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:43Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:43 crc kubenswrapper[4739]: I1201 15:35:43.987569 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:43Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.006437 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"val\\\\nI1201 15:35:30.823819 6160 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:30.823835 6160 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 15:35:30.823846 6160 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 15:35:30.823851 6160 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 15:35:30.823862 6160 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:30.823867 6160 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:30.823878 6160 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 15:35:30.823895 6160 factory.go:656] Stopping watch factory\\\\nI1201 15:35:30.823900 6160 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:30.823895 6160 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:30.823911 6160 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:30.823918 6160 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:30.823926 6160 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:30.823937 6160 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:30.823953 6160 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:30.823961 6160 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:44Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.015877 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:44Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.024854 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:44Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.036925 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:44Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.047277 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.047314 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.047323 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.047339 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.047349 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:44Z","lastTransitionTime":"2025-12-01T15:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.049836 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:44Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.150225 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.150283 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.150301 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.150325 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.150341 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:44Z","lastTransitionTime":"2025-12-01T15:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.252602 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.252663 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.252679 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.252701 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.252718 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:44Z","lastTransitionTime":"2025-12-01T15:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.355417 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.355478 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.355490 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.355507 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.355520 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:44Z","lastTransitionTime":"2025-12-01T15:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.458511 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.458576 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.458599 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.458626 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.458647 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:44Z","lastTransitionTime":"2025-12-01T15:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.561734 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.561798 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.561816 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.561840 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.561858 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:44Z","lastTransitionTime":"2025-12-01T15:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.665007 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.665073 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.665090 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.665113 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.665130 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:44Z","lastTransitionTime":"2025-12-01T15:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.768651 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.768724 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.768745 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.768800 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.768817 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:44Z","lastTransitionTime":"2025-12-01T15:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.796035 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovnkube-controller/2.log" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.797179 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovnkube-controller/1.log" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.800573 4739 generic.go:334] "Generic (PLEG): container finished" podID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerID="a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667" exitCode=1 Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.800630 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerDied","Data":"a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667"} Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.800682 4739 scope.go:117] "RemoveContainer" containerID="796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.801780 4739 scope.go:117] "RemoveContainer" containerID="a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667" Dec 01 15:35:44 crc kubenswrapper[4739]: E1201 15:35:44.802016 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.821846 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:44Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.838509 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:44Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.852206 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:44Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.873058 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.873148 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.873171 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.873204 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.873231 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:44Z","lastTransitionTime":"2025-12-01T15:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.874143 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:44Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.890265 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:44Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.913845 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:44Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.945496 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://796e15da7f26412807955c5f4449a42c80c76d7a3c7a51c01130d5387608fb1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"message\\\":\\\"val\\\\nI1201 15:35:30.823819 6160 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:30.823835 6160 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 15:35:30.823846 6160 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 15:35:30.823851 6160 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 15:35:30.823862 6160 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:30.823867 6160 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:30.823878 6160 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 15:35:30.823895 6160 factory.go:656] Stopping watch factory\\\\nI1201 15:35:30.823900 6160 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:30.823895 6160 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:30.823911 6160 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:30.823918 6160 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:30.823926 6160 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:30.823937 6160 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:30.823953 6160 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:30.823961 6160 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:44Z\\\",\\\"message\\\":\\\"5:44.361286 6372 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 15:35:44.361300 6372 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:44.361320 6372 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:44.361328 6372 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:44.361400 6372 factory.go:656] Stopping watch factory\\\\nI1201 15:35:44.361452 6372 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:44.361492 6372 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:44.361511 6372 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 15:35:44.361523 6372 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:44.361546 6372 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:44.361557 6372 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:44.361566 6372 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:44.361607 6372 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF1201 15:35:44.361661 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to sta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:44Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.963978 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:44Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.978624 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.978662 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.978674 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.978691 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.978704 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:44Z","lastTransitionTime":"2025-12-01T15:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:44 crc kubenswrapper[4739]: I1201 15:35:44.981408 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:44Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.005402 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.026912 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.046440 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.062292 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.079576 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.081090 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.081142 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.081160 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.081184 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.081202 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:45Z","lastTransitionTime":"2025-12-01T15:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.113190 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.132173 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.148071 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.184394 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.184509 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.184528 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.184556 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.184575 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:45Z","lastTransitionTime":"2025-12-01T15:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.287510 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.287571 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.287589 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.287612 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.287630 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:45Z","lastTransitionTime":"2025-12-01T15:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.391393 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.391500 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.391515 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.391540 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.391555 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:45Z","lastTransitionTime":"2025-12-01T15:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.476619 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.476654 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.476699 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.476653 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:45 crc kubenswrapper[4739]: E1201 15:35:45.476859 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:35:45 crc kubenswrapper[4739]: E1201 15:35:45.476986 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:45 crc kubenswrapper[4739]: E1201 15:35:45.477185 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:45 crc kubenswrapper[4739]: E1201 15:35:45.477273 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.495225 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.495282 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.495302 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.495328 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.495348 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:45Z","lastTransitionTime":"2025-12-01T15:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.599547 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.599711 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.599749 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.599779 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.599800 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:45Z","lastTransitionTime":"2025-12-01T15:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.704286 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.704339 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.704355 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.704378 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.704393 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:45Z","lastTransitionTime":"2025-12-01T15:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.806206 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovnkube-controller/2.log" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.806619 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.806653 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.806664 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.806681 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.806692 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:45Z","lastTransitionTime":"2025-12-01T15:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.810531 4739 scope.go:117] "RemoveContainer" containerID="a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667" Dec 01 15:35:45 crc kubenswrapper[4739]: E1201 15:35:45.810689 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.828330 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.841617 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.857002 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.874364 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.896084 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:44Z\\\",\\\"message\\\":\\\"5:44.361286 6372 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 15:35:44.361300 6372 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:44.361320 6372 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:44.361328 6372 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:44.361400 6372 factory.go:656] Stopping watch factory\\\\nI1201 15:35:44.361452 6372 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:44.361492 6372 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:44.361511 6372 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 15:35:44.361523 6372 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:44.361546 6372 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:44.361557 6372 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:44.361566 6372 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:44.361607 6372 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF1201 15:35:44.361661 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to sta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.905579 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.908737 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.908773 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.908781 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.908795 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.908805 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:45Z","lastTransitionTime":"2025-12-01T15:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.914948 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.928391 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.943150 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.955080 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.967981 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.984677 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:45 crc kubenswrapper[4739]: I1201 15:35:45.994168 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:45Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.004745 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:46Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.011139 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.011206 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.011225 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.011253 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.011275 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:46Z","lastTransitionTime":"2025-12-01T15:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.018077 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:46Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.033367 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:46Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.043726 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:46Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.114508 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.114647 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.114667 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.114692 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.114709 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:46Z","lastTransitionTime":"2025-12-01T15:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.219110 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.219181 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.219198 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.219249 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.219267 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:46Z","lastTransitionTime":"2025-12-01T15:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.321921 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.321978 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.322001 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.322027 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.322048 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:46Z","lastTransitionTime":"2025-12-01T15:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.425335 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.425403 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.425463 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.425496 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.425514 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:46Z","lastTransitionTime":"2025-12-01T15:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.528842 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.528913 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.528959 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.528989 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.529010 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:46Z","lastTransitionTime":"2025-12-01T15:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.631734 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.631790 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.631813 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.631841 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.631861 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:46Z","lastTransitionTime":"2025-12-01T15:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.734669 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.734972 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.735154 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.735312 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.735479 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:46Z","lastTransitionTime":"2025-12-01T15:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.838566 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.838629 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.838656 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.838685 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.838704 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:46Z","lastTransitionTime":"2025-12-01T15:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.946164 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.946387 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.946404 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.946457 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:46 crc kubenswrapper[4739]: I1201 15:35:46.946475 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:46Z","lastTransitionTime":"2025-12-01T15:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.049629 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.049703 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.049726 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.049755 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.049774 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:47Z","lastTransitionTime":"2025-12-01T15:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.152718 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.152777 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.152795 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.152818 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.152871 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:47Z","lastTransitionTime":"2025-12-01T15:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.255682 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.255775 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.255801 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.255832 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.255854 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:47Z","lastTransitionTime":"2025-12-01T15:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.358855 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.358907 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.358930 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.358958 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.358978 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:47Z","lastTransitionTime":"2025-12-01T15:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.463782 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.463853 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.463872 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.463900 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.463918 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:47Z","lastTransitionTime":"2025-12-01T15:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.476600 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.476630 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.476600 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.476756 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:47 crc kubenswrapper[4739]: E1201 15:35:47.476929 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:47 crc kubenswrapper[4739]: E1201 15:35:47.477111 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:47 crc kubenswrapper[4739]: E1201 15:35:47.477253 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:47 crc kubenswrapper[4739]: E1201 15:35:47.477384 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.567761 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.567841 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.567878 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.567908 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.567931 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:47Z","lastTransitionTime":"2025-12-01T15:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.670719 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.670769 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.670781 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.670798 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.670810 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:47Z","lastTransitionTime":"2025-12-01T15:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.773691 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.773736 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.773750 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.773770 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.773785 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:47Z","lastTransitionTime":"2025-12-01T15:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.877017 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.877083 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.877099 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.877124 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.877142 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:47Z","lastTransitionTime":"2025-12-01T15:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.980486 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.980571 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.980603 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.980634 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:47 crc kubenswrapper[4739]: I1201 15:35:47.980654 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:47Z","lastTransitionTime":"2025-12-01T15:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.083921 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.083989 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.084007 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.084032 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.084050 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:48Z","lastTransitionTime":"2025-12-01T15:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.187149 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.187210 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.187229 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.187256 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.187274 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:48Z","lastTransitionTime":"2025-12-01T15:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.289726 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.289813 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.289839 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.289874 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.289902 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:48Z","lastTransitionTime":"2025-12-01T15:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.392665 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.392736 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.392760 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.392792 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.392810 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:48Z","lastTransitionTime":"2025-12-01T15:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.495220 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.495278 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.495296 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.495322 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.495340 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:48Z","lastTransitionTime":"2025-12-01T15:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.511334 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.532151 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.550334 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.568292 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.588747 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.598701 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.598739 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.598755 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.598777 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.598796 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:48Z","lastTransitionTime":"2025-12-01T15:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.606030 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.623304 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.640393 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.665153 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:35:48 crc kubenswrapper[4739]: E1201 15:35:48.665367 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:20.665330704 +0000 UTC m=+82.491076838 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.665454 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:48 crc kubenswrapper[4739]: E1201 15:35:48.665685 4739 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:35:48 crc kubenswrapper[4739]: E1201 15:35:48.665772 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:36:20.665750118 +0000 UTC m=+82.491496242 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.666073 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.678165 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.696100 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.701170 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.701225 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.701245 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.701270 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.701289 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:48Z","lastTransitionTime":"2025-12-01T15:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.714568 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.731805 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.750243 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.766087 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.766185 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.766242 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:48 crc kubenswrapper[4739]: E1201 15:35:48.766355 4739 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:35:48 crc kubenswrapper[4739]: E1201 15:35:48.766366 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:35:48 crc kubenswrapper[4739]: E1201 15:35:48.766408 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:35:48 crc kubenswrapper[4739]: E1201 15:35:48.766456 4739 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:48 crc kubenswrapper[4739]: E1201 15:35:48.766470 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:35:48 crc kubenswrapper[4739]: E1201 15:35:48.766515 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:35:48 crc kubenswrapper[4739]: E1201 15:35:48.766489 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:36:20.766449948 +0000 UTC m=+82.592196082 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:35:48 crc kubenswrapper[4739]: E1201 15:35:48.766537 4739 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:48 crc kubenswrapper[4739]: E1201 15:35:48.766564 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 15:36:20.7665395 +0000 UTC m=+82.592285634 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:48 crc kubenswrapper[4739]: E1201 15:35:48.766629 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 15:36:20.766600972 +0000 UTC m=+82.592347106 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.769525 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.799870 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:44Z\\\",\\\"message\\\":\\\"5:44.361286 6372 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 15:35:44.361300 6372 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:44.361320 6372 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:44.361328 6372 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:44.361400 6372 factory.go:656] Stopping watch factory\\\\nI1201 15:35:44.361452 6372 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:44.361492 6372 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:44.361511 6372 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 15:35:44.361523 6372 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:44.361546 6372 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:44.361557 6372 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:44.361566 6372 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:44.361607 6372 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF1201 15:35:44.361661 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to sta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.804017 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.804072 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.804092 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.804116 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.804134 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:48Z","lastTransitionTime":"2025-12-01T15:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.817584 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:48Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.906944 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.907006 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.907024 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.907052 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.907073 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:48Z","lastTransitionTime":"2025-12-01T15:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:48 crc kubenswrapper[4739]: I1201 15:35:48.968273 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs\") pod \"network-metrics-daemon-bsjx4\" (UID: \"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\") " pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:48 crc kubenswrapper[4739]: E1201 15:35:48.968555 4739 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:35:48 crc kubenswrapper[4739]: E1201 15:35:48.968657 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs podName:55ed5d4d-1f46-4e38-9176-2d1a8e5417cf nodeName:}" failed. No retries permitted until 2025-12-01 15:36:04.968626472 +0000 UTC m=+66.794372606 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs") pod "network-metrics-daemon-bsjx4" (UID: "55ed5d4d-1f46-4e38-9176-2d1a8e5417cf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.010504 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.010591 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.010616 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.010648 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.010671 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:49Z","lastTransitionTime":"2025-12-01T15:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.113574 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.113625 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.113639 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.113657 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.113670 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:49Z","lastTransitionTime":"2025-12-01T15:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.217695 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.217800 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.217819 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.217844 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.217862 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:49Z","lastTransitionTime":"2025-12-01T15:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.321470 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.321528 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.321545 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.321567 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.321586 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:49Z","lastTransitionTime":"2025-12-01T15:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.424585 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.424631 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.424647 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.424669 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.424685 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:49Z","lastTransitionTime":"2025-12-01T15:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.527298 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.527344 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.527360 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.527381 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.527397 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:49Z","lastTransitionTime":"2025-12-01T15:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.630581 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.630632 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.630650 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.630674 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.630690 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:49Z","lastTransitionTime":"2025-12-01T15:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.733579 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.733627 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.733644 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.733666 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.733686 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:49Z","lastTransitionTime":"2025-12-01T15:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.836078 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.836122 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.836138 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.836158 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.836174 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:49Z","lastTransitionTime":"2025-12-01T15:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.939116 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.939157 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.939172 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.939194 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:49 crc kubenswrapper[4739]: I1201 15:35:49.939210 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:49Z","lastTransitionTime":"2025-12-01T15:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.042519 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.042570 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.042587 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.042612 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.042629 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:50Z","lastTransitionTime":"2025-12-01T15:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.146268 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.146323 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.146343 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.146367 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.146384 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:50Z","lastTransitionTime":"2025-12-01T15:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.177938 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:50 crc kubenswrapper[4739]: E1201 15:35:50.178119 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.178719 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:50 crc kubenswrapper[4739]: E1201 15:35:50.178850 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.178954 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:50 crc kubenswrapper[4739]: E1201 15:35:50.179078 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.179160 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:50 crc kubenswrapper[4739]: E1201 15:35:50.179239 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.262064 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.262129 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.262149 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.262177 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.262197 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:50Z","lastTransitionTime":"2025-12-01T15:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.367199 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.367288 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.367314 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.367345 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.367369 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:50Z","lastTransitionTime":"2025-12-01T15:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.470643 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.470905 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.470954 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.470986 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.471009 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:50Z","lastTransitionTime":"2025-12-01T15:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.574105 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.574175 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.574199 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.574230 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.574251 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:50Z","lastTransitionTime":"2025-12-01T15:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.677600 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.677664 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.677687 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.677715 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.677736 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:50Z","lastTransitionTime":"2025-12-01T15:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.780714 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.780773 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.780794 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.780824 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.780846 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:50Z","lastTransitionTime":"2025-12-01T15:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.884959 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.885019 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.885036 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.885060 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.885080 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:50Z","lastTransitionTime":"2025-12-01T15:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.987875 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.987937 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.987956 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.987979 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:50 crc kubenswrapper[4739]: I1201 15:35:50.987996 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:50Z","lastTransitionTime":"2025-12-01T15:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.090867 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.090939 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.090964 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.090994 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.091013 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:51Z","lastTransitionTime":"2025-12-01T15:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.194497 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.194595 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.194621 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.194649 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.194672 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:51Z","lastTransitionTime":"2025-12-01T15:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.296861 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.296924 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.296943 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.296968 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.296990 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:51Z","lastTransitionTime":"2025-12-01T15:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.338879 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.356176 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.365339 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.387661 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.399952 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.400033 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.400053 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.400079 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.400101 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:51Z","lastTransitionTime":"2025-12-01T15:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.405972 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.425943 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.442868 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.474311 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:44Z\\\",\\\"message\\\":\\\"5:44.361286 6372 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 15:35:44.361300 6372 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:44.361320 6372 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:44.361328 6372 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:44.361400 6372 factory.go:656] Stopping watch factory\\\\nI1201 15:35:44.361452 6372 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:44.361492 6372 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:44.361511 6372 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 15:35:44.361523 6372 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:44.361546 6372 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:44.361557 6372 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:44.361566 6372 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:44.361607 6372 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF1201 15:35:44.361661 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to sta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.476385 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:51 crc kubenswrapper[4739]: E1201 15:35:51.476617 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.494044 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.503444 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.503505 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.503524 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.503550 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.503568 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:51Z","lastTransitionTime":"2025-12-01T15:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.513342 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.536132 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.558364 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.577658 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.606317 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.606370 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.606386 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.606405 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.606446 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:51Z","lastTransitionTime":"2025-12-01T15:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.611881 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.630584 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.649689 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.665312 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.685283 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.706919 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:51Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.708986 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.709053 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.709068 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.709087 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.709098 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:51Z","lastTransitionTime":"2025-12-01T15:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.812008 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.812077 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.812090 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.812107 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.812120 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:51Z","lastTransitionTime":"2025-12-01T15:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.915230 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.915300 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.915311 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.915328 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:51 crc kubenswrapper[4739]: I1201 15:35:51.915341 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:51Z","lastTransitionTime":"2025-12-01T15:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.018186 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.018297 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.018320 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.018347 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.018369 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:52Z","lastTransitionTime":"2025-12-01T15:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.121627 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.121921 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.122092 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.122277 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.122531 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:52Z","lastTransitionTime":"2025-12-01T15:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.225257 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.225609 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.225812 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.226053 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.226261 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:52Z","lastTransitionTime":"2025-12-01T15:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.329844 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.330139 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.330263 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.330470 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.330625 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:52Z","lastTransitionTime":"2025-12-01T15:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.332393 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.332500 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.332522 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.332546 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.332568 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:52Z","lastTransitionTime":"2025-12-01T15:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:52 crc kubenswrapper[4739]: E1201 15:35:52.352844 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.358100 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.358171 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.358197 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.358229 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.358252 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:52Z","lastTransitionTime":"2025-12-01T15:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:52 crc kubenswrapper[4739]: E1201 15:35:52.378923 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.384559 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.384605 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.384624 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.384647 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.384664 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:52Z","lastTransitionTime":"2025-12-01T15:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:52 crc kubenswrapper[4739]: E1201 15:35:52.406326 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.412086 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.412244 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.412356 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.412518 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.412667 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:52Z","lastTransitionTime":"2025-12-01T15:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:52 crc kubenswrapper[4739]: E1201 15:35:52.431810 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.436368 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.436520 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.436602 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.436688 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.436769 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:52Z","lastTransitionTime":"2025-12-01T15:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:52 crc kubenswrapper[4739]: E1201 15:35:52.454009 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:52Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:52 crc kubenswrapper[4739]: E1201 15:35:52.454333 4739 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.456556 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.456673 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.456762 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.456863 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.456955 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:52Z","lastTransitionTime":"2025-12-01T15:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.476272 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.476317 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:52 crc kubenswrapper[4739]: E1201 15:35:52.476483 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.476519 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:52 crc kubenswrapper[4739]: E1201 15:35:52.476665 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:52 crc kubenswrapper[4739]: E1201 15:35:52.476849 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.560611 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.560667 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.560688 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.560713 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.560732 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:52Z","lastTransitionTime":"2025-12-01T15:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.665403 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.666208 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.666668 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.666891 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.667133 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:52Z","lastTransitionTime":"2025-12-01T15:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.770338 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.770399 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.770461 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.770487 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.770505 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:52Z","lastTransitionTime":"2025-12-01T15:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.874315 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.874381 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.874402 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.874469 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.874488 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:52Z","lastTransitionTime":"2025-12-01T15:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.978385 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.978480 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.978495 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.978515 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:52 crc kubenswrapper[4739]: I1201 15:35:52.978528 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:52Z","lastTransitionTime":"2025-12-01T15:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.081855 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.081943 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.081962 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.081990 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.082009 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:53Z","lastTransitionTime":"2025-12-01T15:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.185089 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.185158 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.185171 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.185187 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.185199 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:53Z","lastTransitionTime":"2025-12-01T15:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.287205 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.287241 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.287252 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.287268 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.287279 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:53Z","lastTransitionTime":"2025-12-01T15:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.390631 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.390775 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.390791 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.390811 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.390864 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:53Z","lastTransitionTime":"2025-12-01T15:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.477107 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:53 crc kubenswrapper[4739]: E1201 15:35:53.477337 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.493574 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.493636 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.493652 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.493675 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.493695 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:53Z","lastTransitionTime":"2025-12-01T15:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.597192 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.597266 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.597290 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.597322 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.597344 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:53Z","lastTransitionTime":"2025-12-01T15:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.700376 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.700409 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.700429 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.700444 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.700453 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:53Z","lastTransitionTime":"2025-12-01T15:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.803133 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.803199 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.803218 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.803241 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.803259 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:53Z","lastTransitionTime":"2025-12-01T15:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.905945 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.906009 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.906026 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.906051 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:53 crc kubenswrapper[4739]: I1201 15:35:53.906144 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:53Z","lastTransitionTime":"2025-12-01T15:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.009675 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.009738 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.009757 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.009783 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.009802 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:54Z","lastTransitionTime":"2025-12-01T15:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.112673 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.112737 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.112754 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.112783 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.112802 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:54Z","lastTransitionTime":"2025-12-01T15:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.216256 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.216385 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.216411 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.216480 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.216506 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:54Z","lastTransitionTime":"2025-12-01T15:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.320493 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.320556 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.320573 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.320598 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.320616 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:54Z","lastTransitionTime":"2025-12-01T15:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.424324 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.424393 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.424455 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.424494 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.424515 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:54Z","lastTransitionTime":"2025-12-01T15:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.476686 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.476706 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.476807 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:54 crc kubenswrapper[4739]: E1201 15:35:54.476966 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:54 crc kubenswrapper[4739]: E1201 15:35:54.477112 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:54 crc kubenswrapper[4739]: E1201 15:35:54.477243 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.527579 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.527638 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.527655 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.527677 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.527694 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:54Z","lastTransitionTime":"2025-12-01T15:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.631187 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.631252 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.631271 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.631298 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.631320 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:54Z","lastTransitionTime":"2025-12-01T15:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.733889 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.733973 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.734015 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.734044 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.734064 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:54Z","lastTransitionTime":"2025-12-01T15:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.836702 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.836768 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.836784 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.836809 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.836826 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:54Z","lastTransitionTime":"2025-12-01T15:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.939928 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.939984 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.939997 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.940014 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:54 crc kubenswrapper[4739]: I1201 15:35:54.940025 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:54Z","lastTransitionTime":"2025-12-01T15:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.042084 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.042119 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.042130 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.042147 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.042158 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:55Z","lastTransitionTime":"2025-12-01T15:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.145524 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.145572 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.145624 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.145650 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.145667 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:55Z","lastTransitionTime":"2025-12-01T15:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.248875 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.249268 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.249469 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.249669 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.249814 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:55Z","lastTransitionTime":"2025-12-01T15:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.352959 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.353013 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.353026 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.353045 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.353056 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:55Z","lastTransitionTime":"2025-12-01T15:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.455898 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.455962 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.455980 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.456005 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.456024 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:55Z","lastTransitionTime":"2025-12-01T15:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.476181 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:55 crc kubenswrapper[4739]: E1201 15:35:55.476360 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.559167 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.559235 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.559254 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.559280 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.559312 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:55Z","lastTransitionTime":"2025-12-01T15:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.663670 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.663721 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.663735 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.663751 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.663762 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:55Z","lastTransitionTime":"2025-12-01T15:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.766843 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.766894 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.766906 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.766922 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.766933 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:55Z","lastTransitionTime":"2025-12-01T15:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.869942 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.869995 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.870008 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.870025 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.870041 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:55Z","lastTransitionTime":"2025-12-01T15:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.974349 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.974406 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.974440 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.974495 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:55 crc kubenswrapper[4739]: I1201 15:35:55.974508 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:55Z","lastTransitionTime":"2025-12-01T15:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.077138 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.077217 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.077235 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.077267 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.077324 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:56Z","lastTransitionTime":"2025-12-01T15:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.180839 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.180921 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.180939 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.180969 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.180993 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:56Z","lastTransitionTime":"2025-12-01T15:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.284982 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.285062 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.285082 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.285111 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.285130 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:56Z","lastTransitionTime":"2025-12-01T15:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.388473 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.388534 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.388551 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.388580 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.388599 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:56Z","lastTransitionTime":"2025-12-01T15:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.477085 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:56 crc kubenswrapper[4739]: E1201 15:35:56.477330 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.477085 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.477456 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:56 crc kubenswrapper[4739]: E1201 15:35:56.477558 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:56 crc kubenswrapper[4739]: E1201 15:35:56.477695 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.491379 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.491458 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.491484 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.491513 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.491532 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:56Z","lastTransitionTime":"2025-12-01T15:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.594773 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.594833 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.594856 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.594900 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.594934 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:56Z","lastTransitionTime":"2025-12-01T15:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.698475 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.698531 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.698548 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.698575 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.698592 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:56Z","lastTransitionTime":"2025-12-01T15:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.801353 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.801457 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.801487 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.801518 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.801543 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:56Z","lastTransitionTime":"2025-12-01T15:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.905283 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.905334 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.905358 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.905381 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:56 crc kubenswrapper[4739]: I1201 15:35:56.905396 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:56Z","lastTransitionTime":"2025-12-01T15:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.008279 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.008327 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.008344 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.008365 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.008381 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:57Z","lastTransitionTime":"2025-12-01T15:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.111578 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.111636 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.111656 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.111684 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.111705 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:57Z","lastTransitionTime":"2025-12-01T15:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.214892 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.215021 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.215042 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.215065 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.215081 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:57Z","lastTransitionTime":"2025-12-01T15:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.318494 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.318555 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.318580 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.318608 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.318629 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:57Z","lastTransitionTime":"2025-12-01T15:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.422191 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.422262 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.422283 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.422316 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.422339 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:57Z","lastTransitionTime":"2025-12-01T15:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.476184 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:57 crc kubenswrapper[4739]: E1201 15:35:57.476363 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.477351 4739 scope.go:117] "RemoveContainer" containerID="a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667" Dec 01 15:35:57 crc kubenswrapper[4739]: E1201 15:35:57.477615 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.525325 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.525375 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.525391 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.525413 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.525457 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:57Z","lastTransitionTime":"2025-12-01T15:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.628726 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.628783 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.628801 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.628823 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.628840 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:57Z","lastTransitionTime":"2025-12-01T15:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.732189 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.732232 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.732242 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.732257 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.732267 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:57Z","lastTransitionTime":"2025-12-01T15:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.834379 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.834462 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.834476 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.834495 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.834507 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:57Z","lastTransitionTime":"2025-12-01T15:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.937609 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.937671 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.937686 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.937711 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:57 crc kubenswrapper[4739]: I1201 15:35:57.937730 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:57Z","lastTransitionTime":"2025-12-01T15:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.040302 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.040365 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.040383 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.040410 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.040478 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:58Z","lastTransitionTime":"2025-12-01T15:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.143756 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.143810 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.143856 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.143885 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.143909 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:58Z","lastTransitionTime":"2025-12-01T15:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.246776 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.246835 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.246847 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.246865 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.246881 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:58Z","lastTransitionTime":"2025-12-01T15:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.349991 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.350074 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.350093 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.350124 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.350148 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:58Z","lastTransitionTime":"2025-12-01T15:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.453919 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.454005 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.454025 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.454055 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.454076 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:58Z","lastTransitionTime":"2025-12-01T15:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.476273 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.476303 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.476344 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:35:58 crc kubenswrapper[4739]: E1201 15:35:58.476541 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:35:58 crc kubenswrapper[4739]: E1201 15:35:58.476674 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:35:58 crc kubenswrapper[4739]: E1201 15:35:58.477017 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.515368 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.536163 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.559351 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.563287 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.563406 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.563474 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.563522 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.563561 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:58Z","lastTransitionTime":"2025-12-01T15:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.585587 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"115b9c1f-3ff6-4919-bfb8-e37c168a7e3d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0adb30e972ba4be0f2150d59de63fc2042ebce799ebc7fe64faa9876ce0009b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d52f33a475d6eb938823fd4b7eac6ca8c026b3826c15bb8b67a50e7a6d942be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da43d8d79a40a66229d0114f6276814c2cdcfbe049eb2a57e89fbb4076154c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.612721 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.633170 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.645861 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.665535 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.669566 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.669619 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.669640 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.669664 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.669685 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:58Z","lastTransitionTime":"2025-12-01T15:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.681103 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.702347 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.719927 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.736408 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.751552 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.768032 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.772402 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.772482 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.772500 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.772552 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.772569 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:58Z","lastTransitionTime":"2025-12-01T15:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.788565 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.810384 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:44Z\\\",\\\"message\\\":\\\"5:44.361286 6372 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 15:35:44.361300 6372 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:44.361320 6372 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:44.361328 6372 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:44.361400 6372 factory.go:656] Stopping watch factory\\\\nI1201 15:35:44.361452 6372 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:44.361492 6372 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:44.361511 6372 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 15:35:44.361523 6372 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:44.361546 6372 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:44.361557 6372 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:44.361566 6372 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:44.361607 6372 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF1201 15:35:44.361661 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to sta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.826562 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.841238 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:35:58Z is after 2025-08-24T17:21:41Z" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.875098 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.875162 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.875185 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.875209 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.875226 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:58Z","lastTransitionTime":"2025-12-01T15:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.977552 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.977599 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.977615 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.977661 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:58 crc kubenswrapper[4739]: I1201 15:35:58.977681 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:58Z","lastTransitionTime":"2025-12-01T15:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.080903 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.080960 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.080977 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.081002 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.081019 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:59Z","lastTransitionTime":"2025-12-01T15:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.183974 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.184041 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.184059 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.184087 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.184104 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:59Z","lastTransitionTime":"2025-12-01T15:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.287630 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.287669 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.287679 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.287695 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.287706 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:59Z","lastTransitionTime":"2025-12-01T15:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.390935 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.390990 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.391003 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.391023 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.391037 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:59Z","lastTransitionTime":"2025-12-01T15:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.476490 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:35:59 crc kubenswrapper[4739]: E1201 15:35:59.476716 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.494041 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.494095 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.494107 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.494126 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.494137 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:59Z","lastTransitionTime":"2025-12-01T15:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.596866 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.596918 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.596930 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.596946 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.596959 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:59Z","lastTransitionTime":"2025-12-01T15:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.699522 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.699582 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.699601 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.699626 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.699644 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:59Z","lastTransitionTime":"2025-12-01T15:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.802950 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.803343 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.803362 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.803387 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.803405 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:59Z","lastTransitionTime":"2025-12-01T15:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.906198 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.906261 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.906274 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.906298 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:35:59 crc kubenswrapper[4739]: I1201 15:35:59.906314 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:35:59Z","lastTransitionTime":"2025-12-01T15:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.009815 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.009859 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.009867 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.009881 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.009889 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:00Z","lastTransitionTime":"2025-12-01T15:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.112377 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.112447 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.112459 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.112476 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.112488 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:00Z","lastTransitionTime":"2025-12-01T15:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.215890 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.215946 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.215965 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.215988 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.216006 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:00Z","lastTransitionTime":"2025-12-01T15:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.318930 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.318971 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.318987 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.319006 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.319021 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:00Z","lastTransitionTime":"2025-12-01T15:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.421575 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.421632 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.421644 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.421665 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.421678 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:00Z","lastTransitionTime":"2025-12-01T15:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.477296 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.477382 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.477337 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:00 crc kubenswrapper[4739]: E1201 15:36:00.477595 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:00 crc kubenswrapper[4739]: E1201 15:36:00.477707 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:00 crc kubenswrapper[4739]: E1201 15:36:00.477818 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.523847 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.523902 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.523914 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.523929 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.523941 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:00Z","lastTransitionTime":"2025-12-01T15:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.626461 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.626572 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.626644 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.626678 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.626703 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:00Z","lastTransitionTime":"2025-12-01T15:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.729156 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.729213 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.729231 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.729257 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.729274 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:00Z","lastTransitionTime":"2025-12-01T15:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.831228 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.831282 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.831300 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.831325 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.831345 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:00Z","lastTransitionTime":"2025-12-01T15:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.934899 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.934940 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.934953 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.934969 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:00 crc kubenswrapper[4739]: I1201 15:36:00.934980 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:00Z","lastTransitionTime":"2025-12-01T15:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.037152 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.037200 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.037216 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.037238 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.037254 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:01Z","lastTransitionTime":"2025-12-01T15:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.141006 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.141077 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.141101 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.141132 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.141152 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:01Z","lastTransitionTime":"2025-12-01T15:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.244348 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.244404 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.244438 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.244462 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.244480 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:01Z","lastTransitionTime":"2025-12-01T15:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.347541 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.347595 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.347612 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.347635 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.347652 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:01Z","lastTransitionTime":"2025-12-01T15:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.450116 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.450158 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.450169 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.450185 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.450196 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:01Z","lastTransitionTime":"2025-12-01T15:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.476847 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:01 crc kubenswrapper[4739]: E1201 15:36:01.477022 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.553227 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.553261 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.553269 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.553283 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.553292 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:01Z","lastTransitionTime":"2025-12-01T15:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.656434 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.656473 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.656486 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.656502 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.656513 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:01Z","lastTransitionTime":"2025-12-01T15:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.759618 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.759698 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.759723 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.759755 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.759779 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:01Z","lastTransitionTime":"2025-12-01T15:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.862682 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.862767 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.862791 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.862824 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.862846 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:01Z","lastTransitionTime":"2025-12-01T15:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.966179 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.966219 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.966227 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.966247 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:01 crc kubenswrapper[4739]: I1201 15:36:01.966255 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:01Z","lastTransitionTime":"2025-12-01T15:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.068847 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.068881 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.068889 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.068902 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.068911 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:02Z","lastTransitionTime":"2025-12-01T15:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.172022 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.172073 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.172111 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.172135 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.172151 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:02Z","lastTransitionTime":"2025-12-01T15:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.274899 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.274944 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.274960 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.274983 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.274999 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:02Z","lastTransitionTime":"2025-12-01T15:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.378025 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.378081 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.378103 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.378131 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.378152 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:02Z","lastTransitionTime":"2025-12-01T15:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.476553 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.476587 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.476656 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:02 crc kubenswrapper[4739]: E1201 15:36:02.476717 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:02 crc kubenswrapper[4739]: E1201 15:36:02.476944 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:02 crc kubenswrapper[4739]: E1201 15:36:02.476845 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.480394 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.480452 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.480461 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.480477 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.480488 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:02Z","lastTransitionTime":"2025-12-01T15:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.582622 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.582679 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.582697 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.582722 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.582741 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:02Z","lastTransitionTime":"2025-12-01T15:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.685587 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.685637 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.685655 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.685679 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.685699 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:02Z","lastTransitionTime":"2025-12-01T15:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.702879 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.702928 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.702945 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.702967 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.702983 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:02Z","lastTransitionTime":"2025-12-01T15:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:02 crc kubenswrapper[4739]: E1201 15:36:02.723370 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.728066 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.728143 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.728168 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.728200 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.728225 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:02Z","lastTransitionTime":"2025-12-01T15:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:02 crc kubenswrapper[4739]: E1201 15:36:02.744867 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.748487 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.748541 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.748555 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.748574 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.748588 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:02Z","lastTransitionTime":"2025-12-01T15:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:02 crc kubenswrapper[4739]: E1201 15:36:02.763896 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.767880 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.767973 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.767993 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.768014 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.768029 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:02Z","lastTransitionTime":"2025-12-01T15:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:02 crc kubenswrapper[4739]: E1201 15:36:02.782521 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.786741 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.786816 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.786839 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.786879 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.786900 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:02Z","lastTransitionTime":"2025-12-01T15:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:02 crc kubenswrapper[4739]: E1201 15:36:02.806126 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:02Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:02 crc kubenswrapper[4739]: E1201 15:36:02.806502 4739 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.808300 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.808361 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.808383 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.808525 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.808553 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:02Z","lastTransitionTime":"2025-12-01T15:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.910590 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.910623 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.910632 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.910645 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:02 crc kubenswrapper[4739]: I1201 15:36:02.910653 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:02Z","lastTransitionTime":"2025-12-01T15:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.013431 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.013483 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.013494 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.013512 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.013524 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:03Z","lastTransitionTime":"2025-12-01T15:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.115311 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.115344 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.115352 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.115366 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.115376 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:03Z","lastTransitionTime":"2025-12-01T15:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.217927 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.217959 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.217970 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.217988 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.218000 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:03Z","lastTransitionTime":"2025-12-01T15:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.320810 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.320844 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.320854 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.320869 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.320880 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:03Z","lastTransitionTime":"2025-12-01T15:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.423791 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.423862 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.423886 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.423916 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.423938 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:03Z","lastTransitionTime":"2025-12-01T15:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.476558 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:03 crc kubenswrapper[4739]: E1201 15:36:03.476706 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.526764 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.526814 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.526825 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.526843 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.526855 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:03Z","lastTransitionTime":"2025-12-01T15:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.628889 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.628940 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.628949 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.628963 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.628972 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:03Z","lastTransitionTime":"2025-12-01T15:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.731144 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.731177 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.731186 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.731200 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.731208 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:03Z","lastTransitionTime":"2025-12-01T15:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.833386 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.833436 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.833444 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.833457 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.833465 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:03Z","lastTransitionTime":"2025-12-01T15:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.934989 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.935018 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.935025 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.935037 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:03 crc kubenswrapper[4739]: I1201 15:36:03.935046 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:03Z","lastTransitionTime":"2025-12-01T15:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.036658 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.036676 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.036685 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.036698 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.036707 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:04Z","lastTransitionTime":"2025-12-01T15:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.138664 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.138719 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.138737 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.138759 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.138775 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:04Z","lastTransitionTime":"2025-12-01T15:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.241092 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.241137 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.241154 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.241176 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.241194 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:04Z","lastTransitionTime":"2025-12-01T15:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.343861 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.343911 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.343924 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.343941 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.343953 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:04Z","lastTransitionTime":"2025-12-01T15:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.446345 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.446371 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.446379 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.446391 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.446399 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:04Z","lastTransitionTime":"2025-12-01T15:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.476809 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.476845 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.476877 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:04 crc kubenswrapper[4739]: E1201 15:36:04.476955 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:04 crc kubenswrapper[4739]: E1201 15:36:04.477101 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:04 crc kubenswrapper[4739]: E1201 15:36:04.477215 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.549073 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.549136 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.549148 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.549161 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.549172 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:04Z","lastTransitionTime":"2025-12-01T15:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.651029 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.651056 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.651064 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.651075 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.651083 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:04Z","lastTransitionTime":"2025-12-01T15:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.753158 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.753200 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.753211 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.753262 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.753274 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:04Z","lastTransitionTime":"2025-12-01T15:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.858965 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.859003 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.859017 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.859033 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.859044 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:04Z","lastTransitionTime":"2025-12-01T15:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.961386 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.961499 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.961518 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.961543 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:04 crc kubenswrapper[4739]: I1201 15:36:04.961561 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:04Z","lastTransitionTime":"2025-12-01T15:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.044158 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs\") pod \"network-metrics-daemon-bsjx4\" (UID: \"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\") " pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:05 crc kubenswrapper[4739]: E1201 15:36:05.044311 4739 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:36:05 crc kubenswrapper[4739]: E1201 15:36:05.044396 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs podName:55ed5d4d-1f46-4e38-9176-2d1a8e5417cf nodeName:}" failed. No retries permitted until 2025-12-01 15:36:37.044375811 +0000 UTC m=+98.870121965 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs") pod "network-metrics-daemon-bsjx4" (UID: "55ed5d4d-1f46-4e38-9176-2d1a8e5417cf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.064097 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.064138 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.064150 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.064164 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.064175 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:05Z","lastTransitionTime":"2025-12-01T15:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.166726 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.166759 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.166769 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.166782 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.166789 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:05Z","lastTransitionTime":"2025-12-01T15:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.269230 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.269259 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.269290 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.269328 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.269337 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:05Z","lastTransitionTime":"2025-12-01T15:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.371065 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.371105 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.371114 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.371128 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.371140 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:05Z","lastTransitionTime":"2025-12-01T15:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.474397 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.474477 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.474491 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.474511 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.474524 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:05Z","lastTransitionTime":"2025-12-01T15:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.476869 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:05 crc kubenswrapper[4739]: E1201 15:36:05.477009 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.577341 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.577401 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.577561 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.577586 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.577598 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:05Z","lastTransitionTime":"2025-12-01T15:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.680196 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.680253 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.680269 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.680291 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.680308 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:05Z","lastTransitionTime":"2025-12-01T15:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.783699 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.783742 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.783759 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.783787 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.783804 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:05Z","lastTransitionTime":"2025-12-01T15:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.882828 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nt6tv_341a0b94-e13d-45cb-8f1c-2c1d1f101272/kube-multus/0.log" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.882930 4739 generic.go:334] "Generic (PLEG): container finished" podID="341a0b94-e13d-45cb-8f1c-2c1d1f101272" containerID="13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46" exitCode=1 Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.882989 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nt6tv" event={"ID":"341a0b94-e13d-45cb-8f1c-2c1d1f101272","Type":"ContainerDied","Data":"13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46"} Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.883829 4739 scope.go:117] "RemoveContainer" containerID="13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.885492 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.885529 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.885542 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.885562 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.885575 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:05Z","lastTransitionTime":"2025-12-01T15:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.903553 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.914888 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.927097 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.937620 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"115b9c1f-3ff6-4919-bfb8-e37c168a7e3d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0adb30e972ba4be0f2150d59de63fc2042ebce799ebc7fe64faa9876ce0009b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d52f33a475d6eb938823fd4b7eac6ca8c026b3826c15bb8b67a50e7a6d942be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da43d8d79a40a66229d0114f6276814c2cdcfbe049eb2a57e89fbb4076154c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.950078 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.962635 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.972167 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.984200 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.987597 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.987627 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.987638 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.987652 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.987663 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:05Z","lastTransitionTime":"2025-12-01T15:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:05 crc kubenswrapper[4739]: I1201 15:36:05.995650 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:05Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.009639 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.023999 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.034354 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.050129 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.063488 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.076445 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.090325 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.090400 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.090449 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.090484 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.090506 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:06Z","lastTransitionTime":"2025-12-01T15:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.091933 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.104889 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:36:05Z\\\",\\\"message\\\":\\\"2025-12-01T15:35:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3\\\\n2025-12-01T15:35:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3 to /host/opt/cni/bin/\\\\n2025-12-01T15:35:20Z [verbose] multus-daemon started\\\\n2025-12-01T15:35:20Z [verbose] Readiness Indicator file check\\\\n2025-12-01T15:36:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.122702 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:44Z\\\",\\\"message\\\":\\\"5:44.361286 6372 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 15:35:44.361300 6372 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:44.361320 6372 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:44.361328 6372 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:44.361400 6372 factory.go:656] Stopping watch factory\\\\nI1201 15:35:44.361452 6372 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:44.361492 6372 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:44.361511 6372 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 15:35:44.361523 6372 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:44.361546 6372 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:44.361557 6372 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:44.361566 6372 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:44.361607 6372 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF1201 15:35:44.361661 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to sta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.193506 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.193546 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.193557 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.193575 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.193589 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:06Z","lastTransitionTime":"2025-12-01T15:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.297255 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.297351 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.297383 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.297455 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.297483 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:06Z","lastTransitionTime":"2025-12-01T15:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.400767 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.400823 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.400840 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.400863 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.400881 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:06Z","lastTransitionTime":"2025-12-01T15:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.477042 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:06 crc kubenswrapper[4739]: E1201 15:36:06.477479 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.477575 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:06 crc kubenswrapper[4739]: E1201 15:36:06.477708 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.477733 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:06 crc kubenswrapper[4739]: E1201 15:36:06.477900 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.503768 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.503821 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.503838 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.503861 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.503878 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:06Z","lastTransitionTime":"2025-12-01T15:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.607336 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.607389 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.607405 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.607451 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.607469 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:06Z","lastTransitionTime":"2025-12-01T15:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.710563 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.710648 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.710672 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.710703 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.710727 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:06Z","lastTransitionTime":"2025-12-01T15:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.813690 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.813743 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.813756 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.813774 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.813786 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:06Z","lastTransitionTime":"2025-12-01T15:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.891432 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nt6tv_341a0b94-e13d-45cb-8f1c-2c1d1f101272/kube-multus/0.log" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.891514 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nt6tv" event={"ID":"341a0b94-e13d-45cb-8f1c-2c1d1f101272","Type":"ContainerStarted","Data":"b580da066f2e8e584d58d7c8982dbc4accfa0bd4149637292b4e8550449452f8"} Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.906522 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"115b9c1f-3ff6-4919-bfb8-e37c168a7e3d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0adb30e972ba4be0f2150d59de63fc2042ebce799ebc7fe64faa9876ce0009b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d52f33a475d6eb938823fd4b7eac6ca8c026b3826c15bb8b67a50e7a6d942be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da43d8d79a40a66229d0114f6276814c2cdcfbe049eb2a57e89fbb4076154c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.917755 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.917786 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.917799 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.917818 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.917832 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:06Z","lastTransitionTime":"2025-12-01T15:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.923021 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.935690 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.947684 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.964065 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.978104 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:06 crc kubenswrapper[4739]: I1201 15:36:06.998785 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:06Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.010710 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:07Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.020315 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.020362 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.020374 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.020394 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.020408 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:07Z","lastTransitionTime":"2025-12-01T15:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.022498 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:07Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.039579 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:07Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.054220 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:07Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.065884 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:07Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.076250 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:07Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.093479 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b580da066f2e8e584d58d7c8982dbc4accfa0bd4149637292b4e8550449452f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:36:05Z\\\",\\\"message\\\":\\\"2025-12-01T15:35:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3\\\\n2025-12-01T15:35:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3 to /host/opt/cni/bin/\\\\n2025-12-01T15:35:20Z [verbose] multus-daemon started\\\\n2025-12-01T15:35:20Z [verbose] Readiness Indicator file check\\\\n2025-12-01T15:36:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:36:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:07Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.122480 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.122507 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.122537 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.122554 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.122566 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:07Z","lastTransitionTime":"2025-12-01T15:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.125852 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:44Z\\\",\\\"message\\\":\\\"5:44.361286 6372 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 15:35:44.361300 6372 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:44.361320 6372 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:44.361328 6372 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:44.361400 6372 factory.go:656] Stopping watch factory\\\\nI1201 15:35:44.361452 6372 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:44.361492 6372 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:44.361511 6372 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 15:35:44.361523 6372 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:44.361546 6372 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:44.361557 6372 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:44.361566 6372 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:44.361607 6372 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF1201 15:35:44.361661 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to sta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:07Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.160529 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:07Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.178701 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:07Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.193152 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:07Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.225044 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.225119 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.225133 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.225152 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.225190 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:07Z","lastTransitionTime":"2025-12-01T15:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.327918 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.327975 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.327996 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.328020 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.328040 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:07Z","lastTransitionTime":"2025-12-01T15:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.430711 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.430760 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.430772 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.430790 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.430802 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:07Z","lastTransitionTime":"2025-12-01T15:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.476588 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:07 crc kubenswrapper[4739]: E1201 15:36:07.476743 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.533609 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.533645 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.533670 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.533748 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.533762 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:07Z","lastTransitionTime":"2025-12-01T15:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.636343 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.636384 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.636396 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.636412 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.636445 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:07Z","lastTransitionTime":"2025-12-01T15:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.738941 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.738993 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.739004 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.739021 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.739032 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:07Z","lastTransitionTime":"2025-12-01T15:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.842306 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.842357 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.842370 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.842388 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.842402 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:07Z","lastTransitionTime":"2025-12-01T15:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.945409 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.945483 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.945498 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.945518 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:07 crc kubenswrapper[4739]: I1201 15:36:07.945530 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:07Z","lastTransitionTime":"2025-12-01T15:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.048802 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.048881 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.048898 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.048928 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.048945 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:08Z","lastTransitionTime":"2025-12-01T15:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.151931 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.152002 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.152026 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.152052 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.152070 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:08Z","lastTransitionTime":"2025-12-01T15:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.255384 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.255454 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.255469 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.255491 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.255504 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:08Z","lastTransitionTime":"2025-12-01T15:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.359323 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.359400 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.359438 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.359465 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.359501 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:08Z","lastTransitionTime":"2025-12-01T15:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.462826 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.462869 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.462883 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.462903 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.462917 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:08Z","lastTransitionTime":"2025-12-01T15:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.477239 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:08 crc kubenswrapper[4739]: E1201 15:36:08.478312 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.478793 4739 scope.go:117] "RemoveContainer" containerID="a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.480953 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:08 crc kubenswrapper[4739]: E1201 15:36:08.481165 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.481715 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:08 crc kubenswrapper[4739]: E1201 15:36:08.481848 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.502009 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b580da066f2e8e584d58d7c8982dbc4accfa0bd4149637292b4e8550449452f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:36:05Z\\\",\\\"message\\\":\\\"2025-12-01T15:35:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3\\\\n2025-12-01T15:35:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3 to /host/opt/cni/bin/\\\\n2025-12-01T15:35:20Z [verbose] multus-daemon started\\\\n2025-12-01T15:35:20Z [verbose] Readiness Indicator file check\\\\n2025-12-01T15:36:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:36:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.536236 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:44Z\\\",\\\"message\\\":\\\"5:44.361286 6372 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 15:35:44.361300 6372 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:44.361320 6372 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:44.361328 6372 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:44.361400 6372 factory.go:656] Stopping watch factory\\\\nI1201 15:35:44.361452 6372 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:44.361492 6372 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:44.361511 6372 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 15:35:44.361523 6372 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:44.361546 6372 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:44.361557 6372 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:44.361566 6372 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:44.361607 6372 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF1201 15:35:44.361661 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to sta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.554764 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.565801 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.565956 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.565984 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.566053 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.566093 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:08Z","lastTransitionTime":"2025-12-01T15:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.570962 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.593385 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.618487 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.636528 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.658393 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.677964 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.677997 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.678007 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.678025 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.678035 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:08Z","lastTransitionTime":"2025-12-01T15:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.680395 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.696235 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.716046 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.730604 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"115b9c1f-3ff6-4919-bfb8-e37c168a7e3d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0adb30e972ba4be0f2150d59de63fc2042ebce799ebc7fe64faa9876ce0009b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d52f33a475d6eb938823fd4b7eac6ca8c026b3826c15bb8b67a50e7a6d942be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da43d8d79a40a66229d0114f6276814c2cdcfbe049eb2a57e89fbb4076154c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.747452 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.766248 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.780635 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.780680 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.780690 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.780704 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.780715 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:08Z","lastTransitionTime":"2025-12-01T15:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.780942 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.796851 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.810370 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.833171 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.882834 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.882863 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.882872 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.882885 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.882894 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:08Z","lastTransitionTime":"2025-12-01T15:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.899703 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovnkube-controller/2.log" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.901905 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerStarted","Data":"e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb"} Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.902602 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.921752 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.945266 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.960755 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.977273 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.985092 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.985130 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.985139 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.985160 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.985172 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:08Z","lastTransitionTime":"2025-12-01T15:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:08 crc kubenswrapper[4739]: I1201 15:36:08.996603 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b580da066f2e8e584d58d7c8982dbc4accfa0bd4149637292b4e8550449452f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:36:05Z\\\",\\\"message\\\":\\\"2025-12-01T15:35:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3\\\\n2025-12-01T15:35:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3 to /host/opt/cni/bin/\\\\n2025-12-01T15:35:20Z [verbose] multus-daemon started\\\\n2025-12-01T15:35:20Z [verbose] Readiness Indicator file check\\\\n2025-12-01T15:36:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:36:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:08Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.012656 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:44Z\\\",\\\"message\\\":\\\"5:44.361286 6372 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 15:35:44.361300 6372 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:44.361320 6372 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:44.361328 6372 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:44.361400 6372 factory.go:656] Stopping watch factory\\\\nI1201 15:35:44.361452 6372 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:44.361492 6372 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:44.361511 6372 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 15:35:44.361523 6372 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:44.361546 6372 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:44.361557 6372 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:44.361566 6372 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:44.361607 6372 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF1201 15:35:44.361661 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to sta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:36:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.023364 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.034374 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.054383 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.066762 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.077691 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.087401 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.087478 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.087489 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.087510 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.087520 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:09Z","lastTransitionTime":"2025-12-01T15:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.089823 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"115b9c1f-3ff6-4919-bfb8-e37c168a7e3d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0adb30e972ba4be0f2150d59de63fc2042ebce799ebc7fe64faa9876ce0009b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d52f33a475d6eb938823fd4b7eac6ca8c026b3826c15bb8b67a50e7a6d942be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da43d8d79a40a66229d0114f6276814c2cdcfbe049eb2a57e89fbb4076154c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.105164 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.124129 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.151607 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.167433 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.180510 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.190326 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.190369 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.190384 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.190403 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.190429 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:09Z","lastTransitionTime":"2025-12-01T15:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.197204 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.292968 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.293006 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.293016 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.293028 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.293038 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:09Z","lastTransitionTime":"2025-12-01T15:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.394957 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.394993 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.395005 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.395020 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.395033 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:09Z","lastTransitionTime":"2025-12-01T15:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.477084 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:09 crc kubenswrapper[4739]: E1201 15:36:09.477304 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.497494 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.497557 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.497574 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.497601 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.497618 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:09Z","lastTransitionTime":"2025-12-01T15:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.600076 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.600120 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.600132 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.600150 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.600164 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:09Z","lastTransitionTime":"2025-12-01T15:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.703189 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.703224 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.703235 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.703251 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.703261 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:09Z","lastTransitionTime":"2025-12-01T15:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.806005 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.806054 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.806070 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.806096 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.806113 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:09Z","lastTransitionTime":"2025-12-01T15:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.907897 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovnkube-controller/3.log" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.908147 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.908217 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.908241 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.908271 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.908294 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:09Z","lastTransitionTime":"2025-12-01T15:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.908879 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovnkube-controller/2.log" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.911820 4739 generic.go:334] "Generic (PLEG): container finished" podID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerID="e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb" exitCode=1 Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.911869 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerDied","Data":"e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb"} Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.911957 4739 scope.go:117] "RemoveContainer" containerID="a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.912481 4739 scope.go:117] "RemoveContainer" containerID="e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb" Dec 01 15:36:09 crc kubenswrapper[4739]: E1201 15:36:09.912644 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.932322 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b580da066f2e8e584d58d7c8982dbc4accfa0bd4149637292b4e8550449452f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:36:05Z\\\",\\\"message\\\":\\\"2025-12-01T15:35:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3\\\\n2025-12-01T15:35:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3 to /host/opt/cni/bin/\\\\n2025-12-01T15:35:20Z [verbose] multus-daemon started\\\\n2025-12-01T15:35:20Z [verbose] Readiness Indicator file check\\\\n2025-12-01T15:36:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:36:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.957297 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a17d0207d9f85360f98930b8f77c92d05ded69f491caf470fe373a171067a667\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:35:44Z\\\",\\\"message\\\":\\\"5:44.361286 6372 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 15:35:44.361300 6372 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 15:35:44.361320 6372 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 15:35:44.361328 6372 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 15:35:44.361400 6372 factory.go:656] Stopping watch factory\\\\nI1201 15:35:44.361452 6372 ovnkube.go:599] Stopped ovnkube\\\\nI1201 15:35:44.361492 6372 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 15:35:44.361511 6372 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 15:35:44.361523 6372 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 15:35:44.361546 6372 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 15:35:44.361557 6372 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1201 15:35:44.361566 6372 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 15:35:44.361590 6372 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 15:35:44.361607 6372 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF1201 15:35:44.361661 6372 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to sta\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:36:09Z\\\",\\\"message\\\":\\\"9861 6697 ovnkube_controller.go:900] Cache entry expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" but failed to find it\\\\nI1201 15:36:09.409871 6697 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI1201 15:36:09.409876 6697 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf after 0 failed attempt(s)\\\\nI1201 15:36:09.409901 6697 default_network_controller.go:776] Recording success event on pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1201 15:36:09.408791 6697 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1201 15:36:09.409932 6697 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1201 15:36:09.409875 6697 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-controller-manager-operator/metrics]} name:Service_openshift-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:36:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.972451 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:09 crc kubenswrapper[4739]: I1201 15:36:09.984527 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:09Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.002392 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.010928 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.010953 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.010963 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.010977 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.010986 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:10Z","lastTransitionTime":"2025-12-01T15:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.016702 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.030155 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.043910 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.066352 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.081496 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.096608 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.110368 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"115b9c1f-3ff6-4919-bfb8-e37c168a7e3d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0adb30e972ba4be0f2150d59de63fc2042ebce799ebc7fe64faa9876ce0009b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d52f33a475d6eb938823fd4b7eac6ca8c026b3826c15bb8b67a50e7a6d942be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da43d8d79a40a66229d0114f6276814c2cdcfbe049eb2a57e89fbb4076154c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.113094 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.113139 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.113151 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.113167 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.113179 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:10Z","lastTransitionTime":"2025-12-01T15:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.126644 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.140869 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.155104 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.174386 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.188437 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.211527 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.215550 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.215596 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.215608 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.215626 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.215642 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:10Z","lastTransitionTime":"2025-12-01T15:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.318818 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.318876 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.318893 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.318917 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.318933 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:10Z","lastTransitionTime":"2025-12-01T15:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.422274 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.422476 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.422539 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.422598 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.422666 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:10Z","lastTransitionTime":"2025-12-01T15:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.476983 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:10 crc kubenswrapper[4739]: E1201 15:36:10.477342 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.477178 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.476992 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:10 crc kubenswrapper[4739]: E1201 15:36:10.477739 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:10 crc kubenswrapper[4739]: E1201 15:36:10.477625 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.524458 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.524499 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.524512 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.524530 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.524543 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:10Z","lastTransitionTime":"2025-12-01T15:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.627259 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.627508 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.627594 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.627683 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.627760 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:10Z","lastTransitionTime":"2025-12-01T15:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.731608 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.731654 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.731669 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.731692 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.731707 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:10Z","lastTransitionTime":"2025-12-01T15:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.834166 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.834204 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.834212 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.834227 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.834237 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:10Z","lastTransitionTime":"2025-12-01T15:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.916965 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovnkube-controller/3.log" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.920060 4739 scope.go:117] "RemoveContainer" containerID="e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb" Dec 01 15:36:10 crc kubenswrapper[4739]: E1201 15:36:10.920251 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.934622 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.936854 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.936890 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.936900 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.936916 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.936926 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:10Z","lastTransitionTime":"2025-12-01T15:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.948322 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.965545 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"115b9c1f-3ff6-4919-bfb8-e37c168a7e3d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0adb30e972ba4be0f2150d59de63fc2042ebce799ebc7fe64faa9876ce0009b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d52f33a475d6eb938823fd4b7eac6ca8c026b3826c15bb8b67a50e7a6d942be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da43d8d79a40a66229d0114f6276814c2cdcfbe049eb2a57e89fbb4076154c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.983234 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:10 crc kubenswrapper[4739]: I1201 15:36:10.996402 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:10Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.015122 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.031825 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.039866 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.039912 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.039921 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.039936 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.039945 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:11Z","lastTransitionTime":"2025-12-01T15:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.048752 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.065549 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.083124 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b580da066f2e8e584d58d7c8982dbc4accfa0bd4149637292b4e8550449452f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:36:05Z\\\",\\\"message\\\":\\\"2025-12-01T15:35:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3\\\\n2025-12-01T15:35:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3 to /host/opt/cni/bin/\\\\n2025-12-01T15:35:20Z [verbose] multus-daemon started\\\\n2025-12-01T15:35:20Z [verbose] Readiness Indicator file check\\\\n2025-12-01T15:36:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:36:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.108047 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:36:09Z\\\",\\\"message\\\":\\\"9861 6697 ovnkube_controller.go:900] Cache entry expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" but failed to find it\\\\nI1201 15:36:09.409871 6697 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI1201 15:36:09.409876 6697 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf after 0 failed attempt(s)\\\\nI1201 15:36:09.409901 6697 default_network_controller.go:776] Recording success event on pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1201 15:36:09.408791 6697 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1201 15:36:09.409932 6697 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1201 15:36:09.409875 6697 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-controller-manager-operator/metrics]} name:Service_openshift-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:36:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.122312 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.133947 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.142772 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.142819 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.142829 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.142852 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.142867 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:11Z","lastTransitionTime":"2025-12-01T15:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.156974 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.174657 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.189924 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.219551 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.235666 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:11Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.245551 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.245610 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.245627 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.245651 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.245670 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:11Z","lastTransitionTime":"2025-12-01T15:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.347980 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.348052 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.348069 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.348090 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.348106 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:11Z","lastTransitionTime":"2025-12-01T15:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.450881 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.450946 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.450987 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.451011 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.451027 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:11Z","lastTransitionTime":"2025-12-01T15:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.476485 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:11 crc kubenswrapper[4739]: E1201 15:36:11.476634 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.553943 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.554000 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.554017 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.554040 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.554055 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:11Z","lastTransitionTime":"2025-12-01T15:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.656548 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.656620 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.656640 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.656665 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.656683 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:11Z","lastTransitionTime":"2025-12-01T15:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.759376 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.759465 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.759484 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.759508 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.759530 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:11Z","lastTransitionTime":"2025-12-01T15:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.862963 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.863006 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.863019 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.863038 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.863050 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:11Z","lastTransitionTime":"2025-12-01T15:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.965965 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.966037 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.966060 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.966088 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:11 crc kubenswrapper[4739]: I1201 15:36:11.966110 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:11Z","lastTransitionTime":"2025-12-01T15:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.069644 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.069687 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.069701 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.069718 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.069730 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:12Z","lastTransitionTime":"2025-12-01T15:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.172831 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.172890 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.172906 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.172930 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.172946 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:12Z","lastTransitionTime":"2025-12-01T15:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.276387 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.276473 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.276492 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.276518 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.276536 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:12Z","lastTransitionTime":"2025-12-01T15:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.379024 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.379082 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.379106 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.379133 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.379154 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:12Z","lastTransitionTime":"2025-12-01T15:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.476541 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.476616 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:12 crc kubenswrapper[4739]: E1201 15:36:12.476700 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.476644 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:12 crc kubenswrapper[4739]: E1201 15:36:12.476825 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:12 crc kubenswrapper[4739]: E1201 15:36:12.476977 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.481947 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.482008 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.482026 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.482240 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.482272 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:12Z","lastTransitionTime":"2025-12-01T15:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.584898 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.584932 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.584941 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.584953 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.584963 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:12Z","lastTransitionTime":"2025-12-01T15:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.687057 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.687121 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.687144 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.687176 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.687200 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:12Z","lastTransitionTime":"2025-12-01T15:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.789711 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.789745 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.789752 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.789763 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.789772 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:12Z","lastTransitionTime":"2025-12-01T15:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.880155 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.880215 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.880233 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.880257 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.880273 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:12Z","lastTransitionTime":"2025-12-01T15:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:12 crc kubenswrapper[4739]: E1201 15:36:12.901462 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:12Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.906235 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.906280 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.906293 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.906311 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.906323 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:12Z","lastTransitionTime":"2025-12-01T15:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:12 crc kubenswrapper[4739]: E1201 15:36:12.924319 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:12Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.928960 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.929013 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.929024 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.929042 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.929055 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:12Z","lastTransitionTime":"2025-12-01T15:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:12 crc kubenswrapper[4739]: E1201 15:36:12.949551 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:12Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.956523 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.956568 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.956578 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.956596 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.956608 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:12Z","lastTransitionTime":"2025-12-01T15:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:12 crc kubenswrapper[4739]: E1201 15:36:12.974315 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:12Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.979055 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.979108 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.979124 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.979145 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.979163 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:12Z","lastTransitionTime":"2025-12-01T15:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:12 crc kubenswrapper[4739]: E1201 15:36:12.991187 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:12Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:12 crc kubenswrapper[4739]: E1201 15:36:12.991348 4739 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.993160 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.993212 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.993228 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.993250 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:12 crc kubenswrapper[4739]: I1201 15:36:12.993264 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:12Z","lastTransitionTime":"2025-12-01T15:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.095690 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.095757 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.095774 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.095796 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.095812 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:13Z","lastTransitionTime":"2025-12-01T15:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.198343 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.198439 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.198455 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.198472 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.198484 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:13Z","lastTransitionTime":"2025-12-01T15:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.309029 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.309095 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.309115 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.309145 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.309168 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:13Z","lastTransitionTime":"2025-12-01T15:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.412157 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.412206 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.412217 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.412235 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.412246 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:13Z","lastTransitionTime":"2025-12-01T15:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.476858 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:13 crc kubenswrapper[4739]: E1201 15:36:13.477040 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.515261 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.515310 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.515327 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.515350 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.515366 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:13Z","lastTransitionTime":"2025-12-01T15:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.618892 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.618947 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.618963 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.618986 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.619002 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:13Z","lastTransitionTime":"2025-12-01T15:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.721855 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.721909 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.721926 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.721948 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.721965 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:13Z","lastTransitionTime":"2025-12-01T15:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.824738 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.824794 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.824811 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.824831 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.824848 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:13Z","lastTransitionTime":"2025-12-01T15:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.928334 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.928390 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.928408 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.928458 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:13 crc kubenswrapper[4739]: I1201 15:36:13.928475 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:13Z","lastTransitionTime":"2025-12-01T15:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.031610 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.031670 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.031687 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.031709 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.031725 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:14Z","lastTransitionTime":"2025-12-01T15:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.134524 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.134620 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.134643 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.134676 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.134698 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:14Z","lastTransitionTime":"2025-12-01T15:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.237454 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.237526 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.237559 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.237591 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.237612 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:14Z","lastTransitionTime":"2025-12-01T15:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.339874 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.339923 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.339936 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.339952 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.339963 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:14Z","lastTransitionTime":"2025-12-01T15:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.442786 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.442849 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.442868 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.442908 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.442926 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:14Z","lastTransitionTime":"2025-12-01T15:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.476415 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.476464 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.476406 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:14 crc kubenswrapper[4739]: E1201 15:36:14.476590 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:14 crc kubenswrapper[4739]: E1201 15:36:14.476649 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:14 crc kubenswrapper[4739]: E1201 15:36:14.476715 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.546017 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.546076 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.546093 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.546118 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.546139 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:14Z","lastTransitionTime":"2025-12-01T15:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.649051 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.649105 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.649124 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.649146 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.649163 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:14Z","lastTransitionTime":"2025-12-01T15:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.752693 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.752758 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.752781 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.752811 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.752834 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:14Z","lastTransitionTime":"2025-12-01T15:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.855696 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.855750 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.855772 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.855804 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.855826 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:14Z","lastTransitionTime":"2025-12-01T15:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.959286 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.959372 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.959396 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.959531 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:14 crc kubenswrapper[4739]: I1201 15:36:14.959578 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:14Z","lastTransitionTime":"2025-12-01T15:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.062205 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.062259 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.062277 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.062299 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.062316 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:15Z","lastTransitionTime":"2025-12-01T15:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.165184 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.165253 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.165271 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.165296 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.165315 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:15Z","lastTransitionTime":"2025-12-01T15:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.268942 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.269001 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.269018 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.269043 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.269060 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:15Z","lastTransitionTime":"2025-12-01T15:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.372446 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.372509 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.372532 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.372561 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.372584 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:15Z","lastTransitionTime":"2025-12-01T15:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.475961 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.476008 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.476028 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.476056 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.476077 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:15Z","lastTransitionTime":"2025-12-01T15:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.476136 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:15 crc kubenswrapper[4739]: E1201 15:36:15.476288 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.579994 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.580065 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.580084 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.580108 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.580126 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:15Z","lastTransitionTime":"2025-12-01T15:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.682949 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.683028 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.683071 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.683096 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.683113 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:15Z","lastTransitionTime":"2025-12-01T15:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.786167 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.786252 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.786275 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.786299 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.786320 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:15Z","lastTransitionTime":"2025-12-01T15:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.889114 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.889180 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.889198 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.889224 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.889242 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:15Z","lastTransitionTime":"2025-12-01T15:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.991577 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.991705 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.991722 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.991748 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:15 crc kubenswrapper[4739]: I1201 15:36:15.991766 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:15Z","lastTransitionTime":"2025-12-01T15:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.094337 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.094403 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.094464 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.094489 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.094507 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:16Z","lastTransitionTime":"2025-12-01T15:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.197288 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.197353 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.197376 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.197403 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.197464 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:16Z","lastTransitionTime":"2025-12-01T15:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.300488 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.300564 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.300588 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.300614 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.300631 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:16Z","lastTransitionTime":"2025-12-01T15:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.404062 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.404130 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.404154 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.404182 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.404204 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:16Z","lastTransitionTime":"2025-12-01T15:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.476470 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.476750 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.476784 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:16 crc kubenswrapper[4739]: E1201 15:36:16.476870 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:16 crc kubenswrapper[4739]: E1201 15:36:16.477050 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:16 crc kubenswrapper[4739]: E1201 15:36:16.477175 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.492998 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.507408 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.507512 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.507536 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.507563 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.507587 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:16Z","lastTransitionTime":"2025-12-01T15:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.610281 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.610329 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.610344 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.610366 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.610383 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:16Z","lastTransitionTime":"2025-12-01T15:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.713143 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.713188 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.713203 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.713224 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.713241 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:16Z","lastTransitionTime":"2025-12-01T15:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.816182 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.816249 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.816274 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.816303 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.816325 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:16Z","lastTransitionTime":"2025-12-01T15:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.919756 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.919843 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.919861 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.919882 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:16 crc kubenswrapper[4739]: I1201 15:36:16.919899 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:16Z","lastTransitionTime":"2025-12-01T15:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.022913 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.022980 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.023006 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.023038 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.023061 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:17Z","lastTransitionTime":"2025-12-01T15:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.125848 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.125920 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.125944 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.125974 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.125998 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:17Z","lastTransitionTime":"2025-12-01T15:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.228535 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.228604 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.228627 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.228651 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.228668 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:17Z","lastTransitionTime":"2025-12-01T15:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.332375 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.332486 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.332510 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.332540 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.332564 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:17Z","lastTransitionTime":"2025-12-01T15:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.435956 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.436023 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.436042 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.436067 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.436086 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:17Z","lastTransitionTime":"2025-12-01T15:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.476916 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:17 crc kubenswrapper[4739]: E1201 15:36:17.477093 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.539043 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.539098 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.539115 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.539140 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.539156 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:17Z","lastTransitionTime":"2025-12-01T15:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.641900 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.641933 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.641942 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.641970 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.641980 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:17Z","lastTransitionTime":"2025-12-01T15:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.744720 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.744774 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.744795 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.744818 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.744838 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:17Z","lastTransitionTime":"2025-12-01T15:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.848509 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.848598 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.848616 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.848640 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.848657 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:17Z","lastTransitionTime":"2025-12-01T15:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.951820 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.951883 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.951906 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.951937 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:17 crc kubenswrapper[4739]: I1201 15:36:17.951961 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:17Z","lastTransitionTime":"2025-12-01T15:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.061654 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.061705 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.061717 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.061775 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.061803 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:18Z","lastTransitionTime":"2025-12-01T15:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.164872 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.164935 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.164952 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.164978 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.164995 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:18Z","lastTransitionTime":"2025-12-01T15:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.268061 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.268122 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.268139 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.268165 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.268183 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:18Z","lastTransitionTime":"2025-12-01T15:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.371948 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.372005 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.372022 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.372045 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.372061 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:18Z","lastTransitionTime":"2025-12-01T15:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.475461 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.475547 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.475574 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.475607 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.475629 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:18Z","lastTransitionTime":"2025-12-01T15:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.476204 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:18 crc kubenswrapper[4739]: E1201 15:36:18.476347 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.476601 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:18 crc kubenswrapper[4739]: E1201 15:36:18.476794 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.477035 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:18 crc kubenswrapper[4739]: E1201 15:36:18.477141 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.497155 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"115b9c1f-3ff6-4919-bfb8-e37c168a7e3d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0adb30e972ba4be0f2150d59de63fc2042ebce799ebc7fe64faa9876ce0009b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d52f33a475d6eb938823fd4b7eac6ca8c026b3826c15bb8b67a50e7a6d942be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da43d8d79a40a66229d0114f6276814c2cdcfbe049eb2a57e89fbb4076154c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.515598 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.534620 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.549035 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.563153 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bfe47b0-4c5d-4246-a435-3a0f7244f300\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://741ce01ef8ea8b4716ee843f9642792ca701fa4e7af0b858817e43f02aa8c6d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b31738b9fc87d330a787b510347c3e96279a085114fccc6d231b87233e7eced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b31738b9fc87d330a787b510347c3e96279a085114fccc6d231b87233e7eced\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.577949 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.578077 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.578163 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.578243 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.578269 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:18Z","lastTransitionTime":"2025-12-01T15:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.578816 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.590261 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.608028 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.635260 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:36:09Z\\\",\\\"message\\\":\\\"9861 6697 ovnkube_controller.go:900] Cache entry expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" but failed to find it\\\\nI1201 15:36:09.409871 6697 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI1201 15:36:09.409876 6697 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf after 0 failed attempt(s)\\\\nI1201 15:36:09.409901 6697 default_network_controller.go:776] Recording success event on pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1201 15:36:09.408791 6697 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1201 15:36:09.409932 6697 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1201 15:36:09.409875 6697 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-controller-manager-operator/metrics]} name:Service_openshift-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:36:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.646970 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.667981 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.681093 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.681172 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.681198 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.681227 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.681248 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:18Z","lastTransitionTime":"2025-12-01T15:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.694158 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.707559 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.725010 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.739409 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.753813 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b580da066f2e8e584d58d7c8982dbc4accfa0bd4149637292b4e8550449452f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:36:05Z\\\",\\\"message\\\":\\\"2025-12-01T15:35:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3\\\\n2025-12-01T15:35:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3 to /host/opt/cni/bin/\\\\n2025-12-01T15:35:20Z [verbose] multus-daemon started\\\\n2025-12-01T15:35:20Z [verbose] Readiness Indicator file check\\\\n2025-12-01T15:36:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:36:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.784141 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.784218 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.784240 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.784270 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.784292 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:18Z","lastTransitionTime":"2025-12-01T15:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.787018 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.805138 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.823824 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:18Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.887068 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.887108 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.887120 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.887135 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.887145 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:18Z","lastTransitionTime":"2025-12-01T15:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.989502 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.989554 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.989571 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.989594 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:18 crc kubenswrapper[4739]: I1201 15:36:18.989611 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:18Z","lastTransitionTime":"2025-12-01T15:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.093092 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.093151 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.093169 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.093192 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.093209 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:19Z","lastTransitionTime":"2025-12-01T15:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.196977 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.197040 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.197056 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.197078 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.197095 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:19Z","lastTransitionTime":"2025-12-01T15:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.301076 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.301134 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.301151 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.301176 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.301194 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:19Z","lastTransitionTime":"2025-12-01T15:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.404041 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.404097 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.404115 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.404142 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.404160 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:19Z","lastTransitionTime":"2025-12-01T15:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.476098 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:19 crc kubenswrapper[4739]: E1201 15:36:19.476264 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.506629 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.506688 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.506708 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.506732 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.506749 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:19Z","lastTransitionTime":"2025-12-01T15:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.609390 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.609481 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.609503 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.609531 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.609552 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:19Z","lastTransitionTime":"2025-12-01T15:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.712820 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.712875 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.712895 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.712920 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.712937 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:19Z","lastTransitionTime":"2025-12-01T15:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.815610 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.815950 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.815968 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.815991 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.816008 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:19Z","lastTransitionTime":"2025-12-01T15:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.919059 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.919108 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.919128 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.919152 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:19 crc kubenswrapper[4739]: I1201 15:36:19.919169 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:19Z","lastTransitionTime":"2025-12-01T15:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.022528 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.022590 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.022605 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.022625 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.022640 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:20Z","lastTransitionTime":"2025-12-01T15:36:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.125647 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.125713 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.125733 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.125765 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.125785 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:20Z","lastTransitionTime":"2025-12-01T15:36:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.228873 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.228930 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.228949 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.228971 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.228987 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:20Z","lastTransitionTime":"2025-12-01T15:36:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.331860 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.331935 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.331960 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.331994 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.332022 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:20Z","lastTransitionTime":"2025-12-01T15:36:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.435051 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.435124 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.435148 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.435182 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.435206 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:20Z","lastTransitionTime":"2025-12-01T15:36:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.476935 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.476973 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.477081 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.477085 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.477235 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.477392 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.538342 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.538388 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.538399 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.538438 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.538451 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:20Z","lastTransitionTime":"2025-12-01T15:36:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.641710 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.641763 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.641774 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.641792 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.641804 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:20Z","lastTransitionTime":"2025-12-01T15:36:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.711924 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.712134 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:24.712100974 +0000 UTC m=+146.537847078 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.712252 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.712446 4739 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.712565 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:37:24.712530337 +0000 UTC m=+146.538276461 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.745155 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.745212 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.745229 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.745253 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.745271 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:20Z","lastTransitionTime":"2025-12-01T15:36:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.813325 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.813458 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.813508 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.813622 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.813652 4739 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.813675 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.813698 4739 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.813703 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.813749 4739 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.813761 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 15:37:24.813726008 +0000 UTC m=+146.639472142 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.813804 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 15:37:24.81378739 +0000 UTC m=+146.639533514 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.813764 4739 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:36:20 crc kubenswrapper[4739]: E1201 15:36:20.813995 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 15:37:24.813950594 +0000 UTC m=+146.639696728 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.848188 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.848265 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.848279 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.848303 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.848329 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:20Z","lastTransitionTime":"2025-12-01T15:36:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.952231 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.952290 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.952310 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.952334 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:20 crc kubenswrapper[4739]: I1201 15:36:20.952351 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:20Z","lastTransitionTime":"2025-12-01T15:36:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.055375 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.055476 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.055496 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.055518 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.055539 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:21Z","lastTransitionTime":"2025-12-01T15:36:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.158316 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.158374 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.158390 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.158414 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.158458 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:21Z","lastTransitionTime":"2025-12-01T15:36:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.260952 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.261033 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.261046 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.261065 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.261114 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:21Z","lastTransitionTime":"2025-12-01T15:36:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.362907 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.362954 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.362970 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.362991 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.363007 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:21Z","lastTransitionTime":"2025-12-01T15:36:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.466623 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.466978 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.467121 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.467303 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.467533 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:21Z","lastTransitionTime":"2025-12-01T15:36:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.477037 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:21 crc kubenswrapper[4739]: E1201 15:36:21.477215 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.571183 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.571253 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.571270 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.571713 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.571783 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:21Z","lastTransitionTime":"2025-12-01T15:36:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.674843 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.674889 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.674906 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.674928 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.674944 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:21Z","lastTransitionTime":"2025-12-01T15:36:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.778373 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.778479 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.778505 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.778536 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.778559 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:21Z","lastTransitionTime":"2025-12-01T15:36:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.882407 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.882491 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.882512 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.882542 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.882560 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:21Z","lastTransitionTime":"2025-12-01T15:36:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.985896 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.985977 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.986001 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.986038 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:21 crc kubenswrapper[4739]: I1201 15:36:21.986062 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:21Z","lastTransitionTime":"2025-12-01T15:36:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.088186 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.088252 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.088277 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.088307 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.088330 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:22Z","lastTransitionTime":"2025-12-01T15:36:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.192522 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.192594 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.192611 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.192681 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.192709 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:22Z","lastTransitionTime":"2025-12-01T15:36:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.296098 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.296147 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.296161 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.296181 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.296196 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:22Z","lastTransitionTime":"2025-12-01T15:36:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.399155 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.399224 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.399248 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.399274 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.399291 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:22Z","lastTransitionTime":"2025-12-01T15:36:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.477178 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.477359 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.477398 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:22 crc kubenswrapper[4739]: E1201 15:36:22.477578 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:22 crc kubenswrapper[4739]: E1201 15:36:22.477685 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:22 crc kubenswrapper[4739]: E1201 15:36:22.477828 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.502258 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.502321 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.502338 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.502362 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.502381 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:22Z","lastTransitionTime":"2025-12-01T15:36:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.604865 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.604916 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.604928 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.604948 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.604960 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:22Z","lastTransitionTime":"2025-12-01T15:36:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.707804 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.707901 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.707920 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.707943 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.707959 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:22Z","lastTransitionTime":"2025-12-01T15:36:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.810859 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.810912 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.810930 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.810955 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.810973 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:22Z","lastTransitionTime":"2025-12-01T15:36:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.914195 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.914252 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.914269 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.914292 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:22 crc kubenswrapper[4739]: I1201 15:36:22.914310 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:22Z","lastTransitionTime":"2025-12-01T15:36:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.016246 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.016307 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.016322 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.016344 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.016356 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:23Z","lastTransitionTime":"2025-12-01T15:36:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.121080 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.121133 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.121147 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.121167 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.121180 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:23Z","lastTransitionTime":"2025-12-01T15:36:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.197106 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.197163 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.197172 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.197191 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.197202 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:23Z","lastTransitionTime":"2025-12-01T15:36:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:23 crc kubenswrapper[4739]: E1201 15:36:23.217763 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.223152 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.223221 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.223244 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.223274 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.223298 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:23Z","lastTransitionTime":"2025-12-01T15:36:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:23 crc kubenswrapper[4739]: E1201 15:36:23.242892 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.252170 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.252209 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.252235 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.252258 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.252269 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:23Z","lastTransitionTime":"2025-12-01T15:36:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:23 crc kubenswrapper[4739]: E1201 15:36:23.274176 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.279271 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.279333 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.279353 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.279379 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.279397 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:23Z","lastTransitionTime":"2025-12-01T15:36:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:23 crc kubenswrapper[4739]: E1201 15:36:23.300680 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.306053 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.306125 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.306143 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.306171 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.306189 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:23Z","lastTransitionTime":"2025-12-01T15:36:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:23 crc kubenswrapper[4739]: E1201 15:36:23.326818 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d50300d8-e6ab-4451-94a5-1b5cdce96a01\\\",\\\"systemUUID\\\":\\\"35294ef1-eb1e-44ab-9a6d-9a0c5248b388\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:23Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:23 crc kubenswrapper[4739]: E1201 15:36:23.327048 4739 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.329497 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.329590 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.329613 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.329636 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.329653 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:23Z","lastTransitionTime":"2025-12-01T15:36:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.432669 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.432734 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.432752 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.432825 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.432847 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:23Z","lastTransitionTime":"2025-12-01T15:36:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.476338 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:23 crc kubenswrapper[4739]: E1201 15:36:23.476683 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.536302 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.536364 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.536383 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.536410 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.536462 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:23Z","lastTransitionTime":"2025-12-01T15:36:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.639464 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.639512 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.639565 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.639593 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.639610 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:23Z","lastTransitionTime":"2025-12-01T15:36:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.742701 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.742783 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.742804 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.742829 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.742845 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:23Z","lastTransitionTime":"2025-12-01T15:36:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.845683 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.845763 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.845776 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.845793 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.845805 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:23Z","lastTransitionTime":"2025-12-01T15:36:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.948330 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.948394 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.948411 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.948465 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:23 crc kubenswrapper[4739]: I1201 15:36:23.948483 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:23Z","lastTransitionTime":"2025-12-01T15:36:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.050952 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.050998 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.051025 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.051041 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.051051 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:24Z","lastTransitionTime":"2025-12-01T15:36:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.153792 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.153820 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.153829 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.153855 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.153869 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:24Z","lastTransitionTime":"2025-12-01T15:36:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.257250 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.257288 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.257300 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.257316 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.257328 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:24Z","lastTransitionTime":"2025-12-01T15:36:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.360782 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.360867 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.360892 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.360922 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.360944 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:24Z","lastTransitionTime":"2025-12-01T15:36:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.465224 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.465292 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.465317 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.465347 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.465365 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:24Z","lastTransitionTime":"2025-12-01T15:36:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.476682 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.476711 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:24 crc kubenswrapper[4739]: E1201 15:36:24.477008 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.477090 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:24 crc kubenswrapper[4739]: E1201 15:36:24.477287 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:24 crc kubenswrapper[4739]: E1201 15:36:24.477558 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.478594 4739 scope.go:117] "RemoveContainer" containerID="e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb" Dec 01 15:36:24 crc kubenswrapper[4739]: E1201 15:36:24.478846 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.568836 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.568900 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.568918 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.568941 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.568959 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:24Z","lastTransitionTime":"2025-12-01T15:36:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.671566 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.671643 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.671667 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.671701 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.671725 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:24Z","lastTransitionTime":"2025-12-01T15:36:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.774631 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.774721 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.774750 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.774780 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.774807 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:24Z","lastTransitionTime":"2025-12-01T15:36:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.878084 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.878165 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.878188 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.878220 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.878243 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:24Z","lastTransitionTime":"2025-12-01T15:36:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.980449 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.980513 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.980530 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.980553 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:24 crc kubenswrapper[4739]: I1201 15:36:24.980571 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:24Z","lastTransitionTime":"2025-12-01T15:36:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.083305 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.083357 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.083373 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.083396 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.083442 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:25Z","lastTransitionTime":"2025-12-01T15:36:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.185669 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.185721 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.185733 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.185753 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.185766 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:25Z","lastTransitionTime":"2025-12-01T15:36:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.288411 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.288498 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.288515 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.288538 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.288552 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:25Z","lastTransitionTime":"2025-12-01T15:36:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.391032 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.391084 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.391102 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.391126 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.391145 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:25Z","lastTransitionTime":"2025-12-01T15:36:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.476892 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:25 crc kubenswrapper[4739]: E1201 15:36:25.477057 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.494282 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.494333 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.494347 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.494367 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.494379 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:25Z","lastTransitionTime":"2025-12-01T15:36:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.597576 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.597681 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.597729 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.597758 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.597774 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:25Z","lastTransitionTime":"2025-12-01T15:36:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.700934 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.701002 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.701021 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.701048 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.701067 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:25Z","lastTransitionTime":"2025-12-01T15:36:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.803586 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.803623 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.803651 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.803665 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.803674 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:25Z","lastTransitionTime":"2025-12-01T15:36:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.906997 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.907077 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.907097 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.907124 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:25 crc kubenswrapper[4739]: I1201 15:36:25.907143 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:25Z","lastTransitionTime":"2025-12-01T15:36:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.009683 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.009752 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.009764 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.009783 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.009796 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:26Z","lastTransitionTime":"2025-12-01T15:36:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.112747 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.112791 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.112803 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.112820 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.112832 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:26Z","lastTransitionTime":"2025-12-01T15:36:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.215736 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.215798 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.215818 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.215842 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.215859 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:26Z","lastTransitionTime":"2025-12-01T15:36:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.319446 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.319510 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.319528 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.319550 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.319568 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:26Z","lastTransitionTime":"2025-12-01T15:36:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.422351 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.422529 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.422555 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.422591 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.422616 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:26Z","lastTransitionTime":"2025-12-01T15:36:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.476949 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.476962 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:26 crc kubenswrapper[4739]: E1201 15:36:26.477785 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:26 crc kubenswrapper[4739]: E1201 15:36:26.477878 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.477064 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:26 crc kubenswrapper[4739]: E1201 15:36:26.478020 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.525244 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.525303 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.525322 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.525345 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.525362 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:26Z","lastTransitionTime":"2025-12-01T15:36:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.628502 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.628569 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.628587 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.628612 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.628631 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:26Z","lastTransitionTime":"2025-12-01T15:36:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.731360 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.731401 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.731413 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.731446 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.731460 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:26Z","lastTransitionTime":"2025-12-01T15:36:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.834639 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.834701 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.834724 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.834752 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.834777 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:26Z","lastTransitionTime":"2025-12-01T15:36:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.937560 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.937606 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.937618 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.937634 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:26 crc kubenswrapper[4739]: I1201 15:36:26.937644 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:26Z","lastTransitionTime":"2025-12-01T15:36:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.040889 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.040944 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.040981 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.041004 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.041021 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:27Z","lastTransitionTime":"2025-12-01T15:36:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.143977 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.144053 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.144087 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.144116 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.144137 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:27Z","lastTransitionTime":"2025-12-01T15:36:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.247453 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.247493 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.247504 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.247520 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.247531 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:27Z","lastTransitionTime":"2025-12-01T15:36:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.350362 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.350472 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.350496 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.350526 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.350551 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:27Z","lastTransitionTime":"2025-12-01T15:36:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.453921 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.453985 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.454019 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.454044 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.454061 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:27Z","lastTransitionTime":"2025-12-01T15:36:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.476235 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:27 crc kubenswrapper[4739]: E1201 15:36:27.476477 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.556643 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.556693 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.557474 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.557518 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.557538 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:27Z","lastTransitionTime":"2025-12-01T15:36:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.661035 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.661130 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.661150 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.661173 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.661191 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:27Z","lastTransitionTime":"2025-12-01T15:36:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.764511 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.764561 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.764577 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.764601 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.764617 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:27Z","lastTransitionTime":"2025-12-01T15:36:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.867943 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.868016 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.868039 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.868068 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.868088 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:27Z","lastTransitionTime":"2025-12-01T15:36:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.970678 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.970958 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.971029 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.971095 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:27 crc kubenswrapper[4739]: I1201 15:36:27.971158 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:27Z","lastTransitionTime":"2025-12-01T15:36:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.074243 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.074305 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.074325 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.074352 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.074412 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:28Z","lastTransitionTime":"2025-12-01T15:36:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.176914 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.177245 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.177409 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.177645 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.177821 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:28Z","lastTransitionTime":"2025-12-01T15:36:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.281700 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.281744 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.281760 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.281782 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.281800 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:28Z","lastTransitionTime":"2025-12-01T15:36:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.384908 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.384999 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.385024 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.385635 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.385679 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:28Z","lastTransitionTime":"2025-12-01T15:36:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.476218 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:28 crc kubenswrapper[4739]: E1201 15:36:28.476411 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.476482 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:28 crc kubenswrapper[4739]: E1201 15:36:28.476739 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.476899 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:28 crc kubenswrapper[4739]: E1201 15:36:28.477051 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.488208 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.488267 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.488286 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.488312 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.488331 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:28Z","lastTransitionTime":"2025-12-01T15:36:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.495728 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.515865 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.537731 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.559935 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nt6tv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"341a0b94-e13d-45cb-8f1c-2c1d1f101272\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:36:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b580da066f2e8e584d58d7c8982dbc4accfa0bd4149637292b4e8550449452f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:36:05Z\\\",\\\"message\\\":\\\"2025-12-01T15:35:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3\\\\n2025-12-01T15:35:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74785cc7-2ab3-46f9-98bd-133f65772ee3 to /host/opt/cni/bin/\\\\n2025-12-01T15:35:20Z [verbose] multus-daemon started\\\\n2025-12-01T15:35:20Z [verbose] Readiness Indicator file check\\\\n2025-12-01T15:36:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:36:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mhdxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nt6tv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.592566 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.592649 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.592675 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.592705 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.592725 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:28Z","lastTransitionTime":"2025-12-01T15:36:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.593274 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a8da1c-7c68-4af9-8b57-817e55d3b875\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T15:36:09Z\\\",\\\"message\\\":\\\"9861 6697 ovnkube_controller.go:900] Cache entry expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" but failed to find it\\\\nI1201 15:36:09.409871 6697 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI1201 15:36:09.409876 6697 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf after 0 failed attempt(s)\\\\nI1201 15:36:09.409901 6697 default_network_controller.go:776] Recording success event on pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1201 15:36:09.408791 6697 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1201 15:36:09.409932 6697 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1201 15:36:09.409875 6697 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-controller-manager-operator/metrics]} name:Service_openshift-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:36:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rvdpq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dlvkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.609030 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e529bfc2eb696d222a0d0ffcbd136e7723d0cf2bb18dc0579c7618f3ba98df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bklr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rnmn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.623043 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5jnk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3613c34f-77d5-414c-b3ae-c31e123ad84b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfb5a64ac4808c602c9d507e211417d2b2d31071a0ba6be316470fb3024b41f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4mjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:21Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5jnk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.644564 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e933a7f-4b33-4d41-bd0e-55f3ee60b933\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T15:35:16Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 15:35:10.791356 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 15:35:10.792464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3507298536/tls.crt::/tmp/serving-cert-3507298536/tls.key\\\\\\\"\\\\nI1201 15:35:16.840245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 15:35:16.843092 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 15:35:16.843126 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 15:35:16.843175 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 15:35:16.843189 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 15:35:16.849856 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 15:35:16.849989 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850015 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 15:35:16.850038 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 15:35:16.850059 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 15:35:16.850079 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 15:35:16.850100 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 15:35:16.849877 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 15:35:16.853647 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.665863 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6ff4b32-f0d5-4db5-a071-01bc01f0ff05\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00ef928b6de932e25563fc3113dc731af68ab4743fb1ce891d2688260f996f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5d9fac00297239de29a054df14ca8e9e8833b3b3fe80785c38f47e85152081\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f3010726316d635644fac71e90f74483f97d0f7b12a26a6e5e8d1ecef6405d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.684959 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6bde8327b5416fd2a1a666c0ee6b43770cee60c7cd912ed4fc6e9320ebc1641e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.695203 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.695277 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.695291 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.695307 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.695340 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:28Z","lastTransitionTime":"2025-12-01T15:36:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.715578 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24344011-0d34-4551-bae4-ae2409751405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78d2f4b1d850c44bffba769dd55ae287e705811f588393c0eb95873c38ac837c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4bd0b022c519845848cffb6c905ca51175d52afe848b1ef15ce52684a295cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff56a82354987e6089bd22ba714f2aaf63ea88b0577119c6e94aeaf72a02dc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd1b300c14d017e45eb30cecb98e46a9cdff819d8bc20c7150170b587e81c59b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe2d8ff91ae2e94513c4aca6b501744d1d6e58faae7593a7396b960d0f49a5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f575701e35d14e40e66f25899c0f50cc150ea2d3e74f3a32ab45e8d1261d589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85c0e5d3f2c9397c373d13d4ac725b6125b9a357d2b0091a1c16e9c55dcf12b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc93cb8260fec7442e4af2231cb6a666003b3917857a0a2eae97a0d19969b631\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.732690 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://013d8ccab8a7cc7e869c6dabb3d044005468aa7026c760421eb2dbfd621b38f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84e7ddff0f04fa6c249bab9b70aaceb02801ae1fc7386b4f82e7549164499324\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.747847 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32ef9ece-a823-470c-8c9c-a3df83efa972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://071f92f3bfa11f14a0287661a765deff8ccd08869ae5bb3779f6d16a60fc0d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7208a8e9221cdf505379983bd8ff85a0d0df08cc40a36fb0129a08b508ea1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xw9z6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ftp8c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.762872 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ts24m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bsjx4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.781577 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"115b9c1f-3ff6-4919-bfb8-e37c168a7e3d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0adb30e972ba4be0f2150d59de63fc2042ebce799ebc7fe64faa9876ce0009b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d52f33a475d6eb938823fd4b7eac6ca8c026b3826c15bb8b67a50e7a6d942be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da43d8d79a40a66229d0114f6276814c2cdcfbe049eb2a57e89fbb4076154c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69c401df3d1d7ffc17c63c9e248ff40c8d28e55e875f7e818864dd8dd416e477\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.797638 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76f50ed0db0f0bfb22218f6ccff354425b943bc7f4af2aafad89d8112e93e883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.797981 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.798077 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.798142 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.798237 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.798335 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:28Z","lastTransitionTime":"2025-12-01T15:36:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.809287 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l4wkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00331ca6-51ac-457e-b95c-9841bda7b582\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61d731956e593adacc6bd824a0ef6d7240b07e6b65631147389b1775bb52f426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8ssk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l4wkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.823115 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vprlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a46c661b-7815-415e-90ac-4be28c3da8bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ff5319a0ef2f4fe9409937f0c41cb157d3e778ef37546f9125aa5c496363343\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:35:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2056e2204de3a2e24f510e0dbc00e5821175d8d5d8145de2bea5144e6b71be3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3decd05b5926c629a64cc3c01177d354ccdc77c8f57e22979c72c589b6c46857\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9eb60f8e615875dac091e0b6da5ad53926643f74177330c1a3b401b1c2997273\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7690c2f0aa8be41613eb5d4f7909187679b1466113ed501c4e3242a246d7c2c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9096214e67381865bf18543c5154ecf113e317f1e1be553b0d54b8e6c26a8bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70ae32847f0e93dbc6b5fd52b04267a044b7dc077355d97636763e5a306800e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:35:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:35:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8bwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:35:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vprlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.837543 4739 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4bfe47b0-4c5d-4246-a435-3a0f7244f300\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T15:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://741ce01ef8ea8b4716ee843f9642792ca701fa4e7af0b858817e43f02aa8c6d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b31738b9fc87d330a787b510347c3e96279a085114fccc6d231b87233e7eced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b31738b9fc87d330a787b510347c3e96279a085114fccc6d231b87233e7eced\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T15:34:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T15:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T15:34:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T15:36:28Z is after 2025-08-24T17:21:41Z" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.901198 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.901291 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.901310 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.901360 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:28 crc kubenswrapper[4739]: I1201 15:36:28.901377 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:28Z","lastTransitionTime":"2025-12-01T15:36:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.004191 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.004245 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.004263 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.004284 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.004302 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:29Z","lastTransitionTime":"2025-12-01T15:36:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.106486 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.106540 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.106556 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.106580 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.106599 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:29Z","lastTransitionTime":"2025-12-01T15:36:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.209766 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.209876 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.209895 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.209922 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.209939 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:29Z","lastTransitionTime":"2025-12-01T15:36:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.312879 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.312964 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.312997 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.313027 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.313056 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:29Z","lastTransitionTime":"2025-12-01T15:36:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.416517 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.416573 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.416596 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.416625 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.416648 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:29Z","lastTransitionTime":"2025-12-01T15:36:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.476713 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:29 crc kubenswrapper[4739]: E1201 15:36:29.476897 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.519547 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.519604 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.519624 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.519646 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.519663 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:29Z","lastTransitionTime":"2025-12-01T15:36:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.622324 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.622389 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.622404 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.622439 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.622456 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:29Z","lastTransitionTime":"2025-12-01T15:36:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.725542 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.725774 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.725791 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.725815 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.725833 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:29Z","lastTransitionTime":"2025-12-01T15:36:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.828040 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.828116 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.828139 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.828169 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.828191 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:29Z","lastTransitionTime":"2025-12-01T15:36:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.931237 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.931289 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.931305 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.931328 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:29 crc kubenswrapper[4739]: I1201 15:36:29.931346 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:29Z","lastTransitionTime":"2025-12-01T15:36:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.034273 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.034353 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.034374 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.034398 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.034455 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:30Z","lastTransitionTime":"2025-12-01T15:36:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.137473 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.137547 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.137570 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.137595 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.137612 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:30Z","lastTransitionTime":"2025-12-01T15:36:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.240703 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.240821 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.240844 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.240875 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.240894 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:30Z","lastTransitionTime":"2025-12-01T15:36:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.344141 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.344230 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.344252 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.344288 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.344312 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:30Z","lastTransitionTime":"2025-12-01T15:36:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.447544 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.447619 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.447647 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.447678 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.447697 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:30Z","lastTransitionTime":"2025-12-01T15:36:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.477274 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.477323 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:30 crc kubenswrapper[4739]: E1201 15:36:30.477468 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.477539 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:30 crc kubenswrapper[4739]: E1201 15:36:30.477677 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:30 crc kubenswrapper[4739]: E1201 15:36:30.477796 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.551362 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.551474 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.551498 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.551530 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.551548 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:30Z","lastTransitionTime":"2025-12-01T15:36:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.653920 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.653971 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.653985 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.654006 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.654022 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:30Z","lastTransitionTime":"2025-12-01T15:36:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.757168 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.757245 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.757270 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.757299 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.757323 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:30Z","lastTransitionTime":"2025-12-01T15:36:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.860863 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.860944 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.860969 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.860998 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.861020 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:30Z","lastTransitionTime":"2025-12-01T15:36:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.963767 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.963833 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.963854 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.963883 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:30 crc kubenswrapper[4739]: I1201 15:36:30.963904 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:30Z","lastTransitionTime":"2025-12-01T15:36:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.066894 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.066996 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.067037 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.067070 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.067092 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:31Z","lastTransitionTime":"2025-12-01T15:36:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.169776 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.169944 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.169968 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.169996 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.170017 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:31Z","lastTransitionTime":"2025-12-01T15:36:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.273104 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.273157 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.273181 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.273207 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.273229 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:31Z","lastTransitionTime":"2025-12-01T15:36:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.376139 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.376222 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.376251 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.376282 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.376306 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:31Z","lastTransitionTime":"2025-12-01T15:36:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.477194 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:31 crc kubenswrapper[4739]: E1201 15:36:31.477949 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.479123 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.479177 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.479198 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.479226 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.479248 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:31Z","lastTransitionTime":"2025-12-01T15:36:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.582380 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.582453 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.582470 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.582495 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.582512 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:31Z","lastTransitionTime":"2025-12-01T15:36:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.685119 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.685168 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.685179 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.685197 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.685210 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:31Z","lastTransitionTime":"2025-12-01T15:36:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.788096 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.788145 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.788158 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.788182 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.788198 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:31Z","lastTransitionTime":"2025-12-01T15:36:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.891657 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.891737 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.891755 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.893061 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.893108 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:31Z","lastTransitionTime":"2025-12-01T15:36:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.995388 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.995442 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.995454 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.995470 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:31 crc kubenswrapper[4739]: I1201 15:36:31.995483 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:31Z","lastTransitionTime":"2025-12-01T15:36:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.098255 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.098356 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.098374 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.098397 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.098445 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:32Z","lastTransitionTime":"2025-12-01T15:36:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.201347 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.201464 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.201483 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.201506 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.201522 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:32Z","lastTransitionTime":"2025-12-01T15:36:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.304746 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.304808 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.304827 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.304851 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.304871 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:32Z","lastTransitionTime":"2025-12-01T15:36:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.407311 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.407376 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.407397 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.407450 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.407473 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:32Z","lastTransitionTime":"2025-12-01T15:36:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.476126 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.476227 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:32 crc kubenswrapper[4739]: E1201 15:36:32.476287 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:32 crc kubenswrapper[4739]: E1201 15:36:32.476400 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.476563 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:32 crc kubenswrapper[4739]: E1201 15:36:32.476820 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.510920 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.510985 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.511009 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.511039 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.511068 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:32Z","lastTransitionTime":"2025-12-01T15:36:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.614024 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.614074 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.614090 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.614115 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.614133 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:32Z","lastTransitionTime":"2025-12-01T15:36:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.718083 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.718148 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.718168 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.718195 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.718211 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:32Z","lastTransitionTime":"2025-12-01T15:36:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.821350 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.821488 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.821508 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.821534 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.821558 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:32Z","lastTransitionTime":"2025-12-01T15:36:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.924337 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.924447 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.924477 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.924507 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:32 crc kubenswrapper[4739]: I1201 15:36:32.924528 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:32Z","lastTransitionTime":"2025-12-01T15:36:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.026318 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.026352 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.026360 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.026373 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.026384 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:33Z","lastTransitionTime":"2025-12-01T15:36:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.129800 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.129861 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.129877 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.129899 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.129917 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:33Z","lastTransitionTime":"2025-12-01T15:36:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.232808 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.232889 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.232916 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.232947 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.232971 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:33Z","lastTransitionTime":"2025-12-01T15:36:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.335964 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.336042 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.336064 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.336093 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.336114 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:33Z","lastTransitionTime":"2025-12-01T15:36:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.338086 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.338139 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.338157 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.338176 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.338193 4739 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T15:36:33Z","lastTransitionTime":"2025-12-01T15:36:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.395764 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668"] Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.396689 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.399243 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.399453 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.400574 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.401037 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.450391 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5v668\" (UID: \"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.450879 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5v668\" (UID: \"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.450920 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5v668\" (UID: \"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.450971 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5v668\" (UID: \"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.451024 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5v668\" (UID: \"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.468222 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=74.468197906 podStartE2EDuration="1m14.468197906s" podCreationTimestamp="2025-12-01 15:35:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:33.445003049 +0000 UTC m=+95.270749183" watchObservedRunningTime="2025-12-01 15:36:33.468197906 +0000 UTC m=+95.293944000" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.476403 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:33 crc kubenswrapper[4739]: E1201 15:36:33.476561 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.484802 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=76.484777498 podStartE2EDuration="1m16.484777498s" podCreationTimestamp="2025-12-01 15:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:33.468707502 +0000 UTC m=+95.294453636" watchObservedRunningTime="2025-12-01 15:36:33.484777498 +0000 UTC m=+95.310523632" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.540763 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=42.540731208 podStartE2EDuration="42.540731208s" podCreationTimestamp="2025-12-01 15:35:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:33.513978135 +0000 UTC m=+95.339724259" watchObservedRunningTime="2025-12-01 15:36:33.540731208 +0000 UTC m=+95.366477352" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.552092 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5v668\" (UID: \"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.552262 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5v668\" (UID: \"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.552330 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5v668\" (UID: \"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.552380 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5v668\" (UID: \"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.552461 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5v668\" (UID: \"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.552625 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5v668\" (UID: \"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.554897 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5v668\" (UID: \"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.554931 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5v668\" (UID: \"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.559660 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5v668\" (UID: \"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.569116 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ftp8c" podStartSLOduration=75.569098559 podStartE2EDuration="1m15.569098559s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:33.568822321 +0000 UTC m=+95.394568425" watchObservedRunningTime="2025-12-01 15:36:33.569098559 +0000 UTC m=+95.394844653" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.589195 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5v668\" (UID: \"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.608926 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-vprlf" podStartSLOduration=75.6089107 podStartE2EDuration="1m15.6089107s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:33.597523662 +0000 UTC m=+95.423269756" watchObservedRunningTime="2025-12-01 15:36:33.6089107 +0000 UTC m=+95.434656794" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.609119 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=17.609115736 podStartE2EDuration="17.609115736s" podCreationTimestamp="2025-12-01 15:36:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:33.608694184 +0000 UTC m=+95.434440288" watchObservedRunningTime="2025-12-01 15:36:33.609115736 +0000 UTC m=+95.434861830" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.658919 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-l4wkn" podStartSLOduration=75.658900563 podStartE2EDuration="1m15.658900563s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:33.641067274 +0000 UTC m=+95.466813408" watchObservedRunningTime="2025-12-01 15:36:33.658900563 +0000 UTC m=+95.484646647" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.676310 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-nt6tv" podStartSLOduration=75.676280198 podStartE2EDuration="1m15.676280198s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:33.675386151 +0000 UTC m=+95.501132255" watchObservedRunningTime="2025-12-01 15:36:33.676280198 +0000 UTC m=+95.502026332" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.718729 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.727699 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podStartSLOduration=75.727668783 podStartE2EDuration="1m15.727668783s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:33.722514029 +0000 UTC m=+95.548260163" watchObservedRunningTime="2025-12-01 15:36:33.727668783 +0000 UTC m=+95.553414917" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.746765 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-5jnk6" podStartSLOduration=75.746746499 podStartE2EDuration="1m15.746746499s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:33.746146791 +0000 UTC m=+95.571892895" watchObservedRunningTime="2025-12-01 15:36:33.746746499 +0000 UTC m=+95.572492593" Dec 01 15:36:33 crc kubenswrapper[4739]: I1201 15:36:33.788373 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=76.788350093 podStartE2EDuration="1m16.788350093s" podCreationTimestamp="2025-12-01 15:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:33.768385 +0000 UTC m=+95.594131104" watchObservedRunningTime="2025-12-01 15:36:33.788350093 +0000 UTC m=+95.614096187" Dec 01 15:36:34 crc kubenswrapper[4739]: I1201 15:36:34.000380 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" event={"ID":"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b","Type":"ContainerStarted","Data":"413a2b77f99e6d2f79ebd5756c3d65135c0f35e06ac725d006f69550d5a25e8a"} Dec 01 15:36:34 crc kubenswrapper[4739]: I1201 15:36:34.000457 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" event={"ID":"70ee0f85-9e77-4cc5-b0dd-e6abd0386c7b","Type":"ContainerStarted","Data":"7f4bfe97b99296b13ada8e46c46154d90649b952dcad99d009532b94c23553d6"} Dec 01 15:36:34 crc kubenswrapper[4739]: I1201 15:36:34.476705 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:34 crc kubenswrapper[4739]: I1201 15:36:34.476789 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:34 crc kubenswrapper[4739]: E1201 15:36:34.476883 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:34 crc kubenswrapper[4739]: I1201 15:36:34.477009 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:34 crc kubenswrapper[4739]: E1201 15:36:34.477101 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:34 crc kubenswrapper[4739]: E1201 15:36:34.477206 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:35 crc kubenswrapper[4739]: I1201 15:36:35.476733 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:35 crc kubenswrapper[4739]: E1201 15:36:35.476916 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:35 crc kubenswrapper[4739]: I1201 15:36:35.477896 4739 scope.go:117] "RemoveContainer" containerID="e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb" Dec 01 15:36:35 crc kubenswrapper[4739]: E1201 15:36:35.478139 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dlvkg_openshift-ovn-kubernetes(43a8da1c-7c68-4af9-8b57-817e55d3b875)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" Dec 01 15:36:36 crc kubenswrapper[4739]: I1201 15:36:36.477179 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:36 crc kubenswrapper[4739]: I1201 15:36:36.477265 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:36 crc kubenswrapper[4739]: E1201 15:36:36.477355 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:36 crc kubenswrapper[4739]: I1201 15:36:36.477477 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:36 crc kubenswrapper[4739]: E1201 15:36:36.477497 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:36 crc kubenswrapper[4739]: E1201 15:36:36.477918 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:37 crc kubenswrapper[4739]: I1201 15:36:37.090655 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs\") pod \"network-metrics-daemon-bsjx4\" (UID: \"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\") " pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:37 crc kubenswrapper[4739]: E1201 15:36:37.090923 4739 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:36:37 crc kubenswrapper[4739]: E1201 15:36:37.091034 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs podName:55ed5d4d-1f46-4e38-9176-2d1a8e5417cf nodeName:}" failed. No retries permitted until 2025-12-01 15:37:41.091004679 +0000 UTC m=+162.916750823 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs") pod "network-metrics-daemon-bsjx4" (UID: "55ed5d4d-1f46-4e38-9176-2d1a8e5417cf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 15:36:37 crc kubenswrapper[4739]: I1201 15:36:37.476873 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:37 crc kubenswrapper[4739]: E1201 15:36:37.477045 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:38 crc kubenswrapper[4739]: I1201 15:36:38.477305 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:38 crc kubenswrapper[4739]: I1201 15:36:38.477331 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:38 crc kubenswrapper[4739]: I1201 15:36:38.477482 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:38 crc kubenswrapper[4739]: E1201 15:36:38.479896 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:38 crc kubenswrapper[4739]: E1201 15:36:38.480076 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:38 crc kubenswrapper[4739]: E1201 15:36:38.480187 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:39 crc kubenswrapper[4739]: I1201 15:36:39.476493 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:39 crc kubenswrapper[4739]: E1201 15:36:39.476625 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:40 crc kubenswrapper[4739]: I1201 15:36:40.476533 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:40 crc kubenswrapper[4739]: I1201 15:36:40.476539 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:40 crc kubenswrapper[4739]: I1201 15:36:40.476651 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:40 crc kubenswrapper[4739]: E1201 15:36:40.476769 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:40 crc kubenswrapper[4739]: E1201 15:36:40.476898 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:40 crc kubenswrapper[4739]: E1201 15:36:40.477026 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:41 crc kubenswrapper[4739]: I1201 15:36:41.476738 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:41 crc kubenswrapper[4739]: E1201 15:36:41.477195 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:42 crc kubenswrapper[4739]: I1201 15:36:42.476264 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:42 crc kubenswrapper[4739]: I1201 15:36:42.476329 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:42 crc kubenswrapper[4739]: I1201 15:36:42.476279 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:42 crc kubenswrapper[4739]: E1201 15:36:42.476489 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:42 crc kubenswrapper[4739]: E1201 15:36:42.476602 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:42 crc kubenswrapper[4739]: E1201 15:36:42.476704 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:43 crc kubenswrapper[4739]: I1201 15:36:43.476810 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:43 crc kubenswrapper[4739]: E1201 15:36:43.476967 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:44 crc kubenswrapper[4739]: I1201 15:36:44.476987 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:44 crc kubenswrapper[4739]: I1201 15:36:44.477070 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:44 crc kubenswrapper[4739]: E1201 15:36:44.477510 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:44 crc kubenswrapper[4739]: I1201 15:36:44.477158 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:44 crc kubenswrapper[4739]: E1201 15:36:44.477778 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:44 crc kubenswrapper[4739]: E1201 15:36:44.477857 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:45 crc kubenswrapper[4739]: I1201 15:36:45.476517 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:45 crc kubenswrapper[4739]: E1201 15:36:45.476694 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:46 crc kubenswrapper[4739]: I1201 15:36:46.476649 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:46 crc kubenswrapper[4739]: I1201 15:36:46.476724 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:46 crc kubenswrapper[4739]: I1201 15:36:46.476649 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:46 crc kubenswrapper[4739]: E1201 15:36:46.476906 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:46 crc kubenswrapper[4739]: E1201 15:36:46.477098 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:46 crc kubenswrapper[4739]: E1201 15:36:46.477196 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:47 crc kubenswrapper[4739]: I1201 15:36:47.477053 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:47 crc kubenswrapper[4739]: E1201 15:36:47.477268 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:48 crc kubenswrapper[4739]: I1201 15:36:48.476201 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:48 crc kubenswrapper[4739]: I1201 15:36:48.476310 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:48 crc kubenswrapper[4739]: I1201 15:36:48.476399 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:48 crc kubenswrapper[4739]: E1201 15:36:48.478120 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:48 crc kubenswrapper[4739]: E1201 15:36:48.478926 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:48 crc kubenswrapper[4739]: E1201 15:36:48.479016 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:49 crc kubenswrapper[4739]: I1201 15:36:49.476768 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:49 crc kubenswrapper[4739]: E1201 15:36:49.477124 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:49 crc kubenswrapper[4739]: I1201 15:36:49.477319 4739 scope.go:117] "RemoveContainer" containerID="e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb" Dec 01 15:36:50 crc kubenswrapper[4739]: I1201 15:36:50.059653 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovnkube-controller/3.log" Dec 01 15:36:50 crc kubenswrapper[4739]: I1201 15:36:50.062687 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerStarted","Data":"edcd593134ac40a8f73daf48b1614b2fe6e7d0f598dd6cd8b741934e8dfbb1dc"} Dec 01 15:36:50 crc kubenswrapper[4739]: I1201 15:36:50.063273 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:36:50 crc kubenswrapper[4739]: I1201 15:36:50.092850 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5v668" podStartSLOduration=92.092828657 podStartE2EDuration="1m32.092828657s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:34.015906232 +0000 UTC m=+95.841652366" watchObservedRunningTime="2025-12-01 15:36:50.092828657 +0000 UTC m=+111.918574781" Dec 01 15:36:50 crc kubenswrapper[4739]: I1201 15:36:50.093228 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podStartSLOduration=92.093220278 podStartE2EDuration="1m32.093220278s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:50.092925229 +0000 UTC m=+111.918671323" watchObservedRunningTime="2025-12-01 15:36:50.093220278 +0000 UTC m=+111.918966402" Dec 01 15:36:50 crc kubenswrapper[4739]: I1201 15:36:50.423168 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-bsjx4"] Dec 01 15:36:50 crc kubenswrapper[4739]: I1201 15:36:50.423472 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:50 crc kubenswrapper[4739]: E1201 15:36:50.423676 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:50 crc kubenswrapper[4739]: I1201 15:36:50.476619 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:50 crc kubenswrapper[4739]: E1201 15:36:50.476880 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:50 crc kubenswrapper[4739]: I1201 15:36:50.477589 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:50 crc kubenswrapper[4739]: E1201 15:36:50.477801 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:51 crc kubenswrapper[4739]: I1201 15:36:51.476786 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:51 crc kubenswrapper[4739]: E1201 15:36:51.477358 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 15:36:52 crc kubenswrapper[4739]: I1201 15:36:52.476317 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:52 crc kubenswrapper[4739]: I1201 15:36:52.476349 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:52 crc kubenswrapper[4739]: E1201 15:36:52.476658 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 15:36:52 crc kubenswrapper[4739]: E1201 15:36:52.477285 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 15:36:52 crc kubenswrapper[4739]: I1201 15:36:52.478139 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:52 crc kubenswrapper[4739]: E1201 15:36:52.478413 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bsjx4" podUID="55ed5d4d-1f46-4e38-9176-2d1a8e5417cf" Dec 01 15:36:52 crc kubenswrapper[4739]: I1201 15:36:52.962740 4739 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 01 15:36:52 crc kubenswrapper[4739]: I1201 15:36:52.962976 4739 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.024713 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j47hk"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.033180 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.035251 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.051241 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.051555 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.051792 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.052135 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.053161 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.053823 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.054280 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.054722 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.056985 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-twbgk"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.057810 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-twbgk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.059941 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-w9l4v"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.060779 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-b9xb8"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.061052 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.061257 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.061635 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mjqqv"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.062306 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.067699 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cs6j4"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.068364 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.076050 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.077398 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-s2hr5"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.078192 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.078802 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.079243 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.080141 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.084261 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-s6nss"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.084501 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.084895 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-5dxtl"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.085385 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-s6nss" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.087230 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.087550 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.095558 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j47hk"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.095688 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-zf9xg"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.096147 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.096227 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.099190 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.105790 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.106380 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-6b8xv"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.106825 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.107125 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.107465 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.108595 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.109105 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.112948 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-sqfdh"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.114007 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.114469 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-sqfdh" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.118150 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kqk8p"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.118902 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.119387 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.119392 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.120722 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.120856 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.127525 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93126f7a-dfe3-4864-bc30-ac3629eb2bc2-config\") pod \"machine-api-operator-5694c8668f-j47hk\" (UID: \"93126f7a-dfe3-4864-bc30-ac3629eb2bc2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.127667 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/93126f7a-dfe3-4864-bc30-ac3629eb2bc2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j47hk\" (UID: \"93126f7a-dfe3-4864-bc30-ac3629eb2bc2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.127699 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpdjx\" (UniqueName: \"kubernetes.io/projected/93126f7a-dfe3-4864-bc30-ac3629eb2bc2-kube-api-access-vpdjx\") pod \"machine-api-operator-5694c8668f-j47hk\" (UID: \"93126f7a-dfe3-4864-bc30-ac3629eb2bc2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.127959 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/93126f7a-dfe3-4864-bc30-ac3629eb2bc2-images\") pod \"machine-api-operator-5694c8668f-j47hk\" (UID: \"93126f7a-dfe3-4864-bc30-ac3629eb2bc2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.130769 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.132085 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.137066 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.143113 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l7mfx"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.144017 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.145379 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.146296 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.148135 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.168599 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.168890 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.169210 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.172641 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.172962 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.174425 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.174648 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.176795 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.177109 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.177364 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.177524 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.177696 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.177939 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.178058 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.180051 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2jcd2"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.180816 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2jcd2" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.182042 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.186898 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-v2m45"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.187599 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-v2m45" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.195326 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.196899 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-b9xb8"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.197060 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.197181 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.197284 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.197319 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.197371 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.198969 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.199177 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.199303 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.199495 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.199616 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.199752 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.199982 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.200116 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.200202 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.200276 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.200346 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.200458 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.200697 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.200769 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.200833 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.200907 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.200977 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.201050 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.201184 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.201249 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.201287 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.201367 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.201417 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.201459 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.201540 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.201590 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.201660 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.201736 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.201881 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.202005 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.202094 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.202167 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.202241 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.201180 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.202542 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.202677 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.202784 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.202800 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.202883 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.202907 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.202958 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.203030 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.203041 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.203058 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.203106 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.203146 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.203171 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.203216 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.203218 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.203411 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.203501 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.206495 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.206633 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.206718 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.206770 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.206798 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.201543 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.206834 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.206877 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.206980 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.207295 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.207451 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.207663 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.213890 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.214715 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.214949 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.215065 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.215835 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.216028 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.216042 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.216043 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.216153 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.218386 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.218541 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.231816 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232406 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz29k\" (UniqueName: \"kubernetes.io/projected/80a392a8-e94f-4dfa-aefa-3dd05e1051c5-kube-api-access-lz29k\") pod \"cluster-samples-operator-665b6dd947-twbgk\" (UID: \"80a392a8-e94f-4dfa-aefa-3dd05e1051c5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-twbgk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232469 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvvck\" (UniqueName: \"kubernetes.io/projected/29e4371e-f34c-409b-b9db-64de91ddd2ef-kube-api-access-wvvck\") pod \"cluster-image-registry-operator-dc59b4c8b-48dh8\" (UID: \"29e4371e-f34c-409b-b9db-64de91ddd2ef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232492 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/68db36cd-270c-4550-8642-44f72796a198-etcd-client\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232512 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-oauth-serving-cert\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232530 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e9c01d1-d2a8-4d89-afb9-9599728cb909-serving-cert\") pod \"console-operator-58897d9998-s2hr5\" (UID: \"8e9c01d1-d2a8-4d89-afb9-9599728cb909\") " pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232548 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/80a392a8-e94f-4dfa-aefa-3dd05e1051c5-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-twbgk\" (UID: \"80a392a8-e94f-4dfa-aefa-3dd05e1051c5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-twbgk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232568 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/217d2158-0c6e-48d9-9713-5499a54655d9-metrics-certs\") pod \"router-default-5444994796-6b8xv\" (UID: \"217d2158-0c6e-48d9-9713-5499a54655d9\") " pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232595 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/93126f7a-dfe3-4864-bc30-ac3629eb2bc2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j47hk\" (UID: \"93126f7a-dfe3-4864-bc30-ac3629eb2bc2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232618 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2bbf\" (UniqueName: \"kubernetes.io/projected/8e9c01d1-d2a8-4d89-afb9-9599728cb909-kube-api-access-d2bbf\") pod \"console-operator-58897d9998-s2hr5\" (UID: \"8e9c01d1-d2a8-4d89-afb9-9599728cb909\") " pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232636 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a-config\") pod \"authentication-operator-69f744f599-cs6j4\" (UID: \"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232657 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42bba742-38b4-4d27-9de8-d5f23fe28e84-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-85h9w\" (UID: \"42bba742-38b4-4d27-9de8-d5f23fe28e84\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232678 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68db36cd-270c-4550-8642-44f72796a198-serving-cert\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232701 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzq7j\" (UniqueName: \"kubernetes.io/projected/217d2158-0c6e-48d9-9713-5499a54655d9-kube-api-access-zzq7j\") pod \"router-default-5444994796-6b8xv\" (UID: \"217d2158-0c6e-48d9-9713-5499a54655d9\") " pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232719 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-l7mfx\" (UID: \"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175\") " pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232738 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6d0c9b6a-66ba-4667-a6ea-833b12934c09-proxy-tls\") pod \"machine-config-operator-74547568cd-8jkjv\" (UID: \"6d0c9b6a-66ba-4667-a6ea-833b12934c09\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232757 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b-proxy-tls\") pod \"machine-config-controller-84d6567774-qvnkw\" (UID: \"7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232775 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-audit-dir\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232794 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/68db36cd-270c-4550-8642-44f72796a198-etcd-serving-ca\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232811 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a-service-ca-bundle\") pod \"authentication-operator-69f744f599-cs6j4\" (UID: \"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232831 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42bba742-38b4-4d27-9de8-d5f23fe28e84-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-85h9w\" (UID: \"42bba742-38b4-4d27-9de8-d5f23fe28e84\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232853 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cs6j4\" (UID: \"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232874 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-audit-policies\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232890 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.232916 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-encryption-config\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233263 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cec3a69-7dbc-4781-a800-452a7e740adc-serving-cert\") pod \"controller-manager-879f6c89f-b9xb8\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233297 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-audit-dir\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233318 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vklx7\" (UniqueName: \"kubernetes.io/projected/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-kube-api-access-vklx7\") pod \"marketplace-operator-79b997595-l7mfx\" (UID: \"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175\") " pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233346 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/93126f7a-dfe3-4864-bc30-ac3629eb2bc2-images\") pod \"machine-api-operator-5694c8668f-j47hk\" (UID: \"93126f7a-dfe3-4864-bc30-ac3629eb2bc2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233366 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93126f7a-dfe3-4864-bc30-ac3629eb2bc2-config\") pod \"machine-api-operator-5694c8668f-j47hk\" (UID: \"93126f7a-dfe3-4864-bc30-ac3629eb2bc2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233385 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29e4371e-f34c-409b-b9db-64de91ddd2ef-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-48dh8\" (UID: \"29e4371e-f34c-409b-b9db-64de91ddd2ef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233403 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/29e4371e-f34c-409b-b9db-64de91ddd2ef-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-48dh8\" (UID: \"29e4371e-f34c-409b-b9db-64de91ddd2ef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233445 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b-config\") pod \"kube-controller-manager-operator-78b949d7b-ndt7g\" (UID: \"12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233466 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/217d2158-0c6e-48d9-9713-5499a54655d9-default-certificate\") pod \"router-default-5444994796-6b8xv\" (UID: \"217d2158-0c6e-48d9-9713-5499a54655d9\") " pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233484 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/68db36cd-270c-4550-8642-44f72796a198-audit\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233501 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zddvn\" (UniqueName: \"kubernetes.io/projected/6d0c9b6a-66ba-4667-a6ea-833b12934c09-kube-api-access-zddvn\") pod \"machine-config-operator-74547568cd-8jkjv\" (UID: \"6d0c9b6a-66ba-4667-a6ea-833b12934c09\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233524 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqmqh\" (UniqueName: \"kubernetes.io/projected/1878674e-9147-4b42-b95a-6c5a82ff5573-kube-api-access-nqmqh\") pod \"openshift-apiserver-operator-796bbdcf4f-6bt2z\" (UID: \"1878674e-9147-4b42-b95a-6c5a82ff5573\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233541 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/68db36cd-270c-4550-8642-44f72796a198-encryption-config\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233559 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233574 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-audit-policies\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233592 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0234bac5-34c7-4ef3-93f2-4673ebf78d7a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bb5fl\" (UID: \"0234bac5-34c7-4ef3-93f2-4673ebf78d7a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233612 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/351ebc2b-c707-4aec-8f6b-e79af328fdf5-config\") pod \"machine-approver-56656f9798-mnwj9\" (UID: \"351ebc2b-c707-4aec-8f6b-e79af328fdf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233630 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233650 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t66tc\" (UniqueName: \"kubernetes.io/projected/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-kube-api-access-t66tc\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233676 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtmhz\" (UniqueName: \"kubernetes.io/projected/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-kube-api-access-wtmhz\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233693 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f97c9c68-21cb-4f92-b725-3130f00005c8-etcd-client\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233709 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c886dc3f-4a59-4e31-a900-2c000427c7a9-client-ca\") pod \"route-controller-manager-6576b87f9c-5pnp5\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233726 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233744 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-client-ca\") pod \"controller-manager-879f6c89f-b9xb8\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233765 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/351ebc2b-c707-4aec-8f6b-e79af328fdf5-machine-approver-tls\") pod \"machine-approver-56656f9798-mnwj9\" (UID: \"351ebc2b-c707-4aec-8f6b-e79af328fdf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233782 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/68d2f0b1-e6a9-4133-aa4d-19c40db638d2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9zvgq\" (UID: \"68d2f0b1-e6a9-4133-aa4d-19c40db638d2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233798 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-ndt7g\" (UID: \"12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233817 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/68db36cd-270c-4550-8642-44f72796a198-trusted-ca-bundle\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233836 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/29e4371e-f34c-409b-b9db-64de91ddd2ef-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-48dh8\" (UID: \"29e4371e-f34c-409b-b9db-64de91ddd2ef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233855 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-ndt7g\" (UID: \"12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233871 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233890 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4kd7\" (UniqueName: \"kubernetes.io/projected/b2995c47-f228-411e-857d-c2830c577ffc-kube-api-access-g4kd7\") pod \"dns-operator-744455d44c-sqfdh\" (UID: \"b2995c47-f228-411e-857d-c2830c577ffc\") " pod="openshift-dns-operator/dns-operator-744455d44c-sqfdh" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233912 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-etcd-client\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233930 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/217d2158-0c6e-48d9-9713-5499a54655d9-stats-auth\") pod \"router-default-5444994796-6b8xv\" (UID: \"217d2158-0c6e-48d9-9713-5499a54655d9\") " pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233947 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-b9xb8\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233962 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlnxt\" (UniqueName: \"kubernetes.io/projected/7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b-kube-api-access-dlnxt\") pod \"machine-config-controller-84d6567774-qvnkw\" (UID: \"7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233985 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6d0c9b6a-66ba-4667-a6ea-833b12934c09-images\") pod \"machine-config-operator-74547568cd-8jkjv\" (UID: \"6d0c9b6a-66ba-4667-a6ea-833b12934c09\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.234038 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2ljj\" (UniqueName: \"kubernetes.io/projected/f97c9c68-21cb-4f92-b725-3130f00005c8-kube-api-access-d2ljj\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.234158 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/351ebc2b-c707-4aec-8f6b-e79af328fdf5-auth-proxy-config\") pod \"machine-approver-56656f9798-mnwj9\" (UID: \"351ebc2b-c707-4aec-8f6b-e79af328fdf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.234185 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qvnkw\" (UID: \"7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.234224 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.234306 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8e9c01d1-d2a8-4d89-afb9-9599728cb909-trusted-ca\") pod \"console-operator-58897d9998-s2hr5\" (UID: \"8e9c01d1-d2a8-4d89-afb9-9599728cb909\") " pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.234326 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.233270 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.234946 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.235255 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93126f7a-dfe3-4864-bc30-ac3629eb2bc2-config\") pod \"machine-api-operator-5694c8668f-j47hk\" (UID: \"93126f7a-dfe3-4864-bc30-ac3629eb2bc2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.234986 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b2995c47-f228-411e-857d-c2830c577ffc-metrics-tls\") pod \"dns-operator-744455d44c-sqfdh\" (UID: \"b2995c47-f228-411e-857d-c2830c577ffc\") " pod="openshift-dns-operator/dns-operator-744455d44c-sqfdh" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.235543 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpdjx\" (UniqueName: \"kubernetes.io/projected/93126f7a-dfe3-4864-bc30-ac3629eb2bc2-kube-api-access-vpdjx\") pod \"machine-api-operator-5694c8668f-j47hk\" (UID: \"93126f7a-dfe3-4864-bc30-ac3629eb2bc2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.235599 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-config\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.235624 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/217d2158-0c6e-48d9-9713-5499a54655d9-service-ca-bundle\") pod \"router-default-5444994796-6b8xv\" (UID: \"217d2158-0c6e-48d9-9713-5499a54655d9\") " pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.237253 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/93126f7a-dfe3-4864-bc30-ac3629eb2bc2-images\") pod \"machine-api-operator-5694c8668f-j47hk\" (UID: \"93126f7a-dfe3-4864-bc30-ac3629eb2bc2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.237799 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.237989 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/38613c37-a98c-443b-8aaf-c75fcabc9238-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-pmd7x\" (UID: \"38613c37-a98c-443b-8aaf-c75fcabc9238\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238036 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0234bac5-34c7-4ef3-93f2-4673ebf78d7a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bb5fl\" (UID: \"0234bac5-34c7-4ef3-93f2-4673ebf78d7a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238050 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238060 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6d0c9b6a-66ba-4667-a6ea-833b12934c09-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8jkjv\" (UID: \"6d0c9b6a-66ba-4667-a6ea-833b12934c09\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238272 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2720c151-cb93-4622-ad01-379f5f4b0466-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-h7kkm\" (UID: \"2720c151-cb93-4622-ad01-379f5f4b0466\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238292 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38613c37-a98c-443b-8aaf-c75fcabc9238-config\") pod \"kube-apiserver-operator-766d6c64bb-pmd7x\" (UID: \"38613c37-a98c-443b-8aaf-c75fcabc9238\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238308 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48x5l\" (UniqueName: \"kubernetes.io/projected/351ebc2b-c707-4aec-8f6b-e79af328fdf5-kube-api-access-48x5l\") pod \"machine-approver-56656f9798-mnwj9\" (UID: \"351ebc2b-c707-4aec-8f6b-e79af328fdf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238329 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-serving-cert\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238348 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1878674e-9147-4b42-b95a-6c5a82ff5573-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6bt2z\" (UID: \"1878674e-9147-4b42-b95a-6c5a82ff5573\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238368 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk5xf\" (UniqueName: \"kubernetes.io/projected/67608617-45fd-4c45-87f9-325996855e5e-kube-api-access-lk5xf\") pod \"downloads-7954f5f757-s6nss\" (UID: \"67608617-45fd-4c45-87f9-325996855e5e\") " pod="openshift-console/downloads-7954f5f757-s6nss" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238405 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtz4r\" (UniqueName: \"kubernetes.io/projected/2720c151-cb93-4622-ad01-379f5f4b0466-kube-api-access-mtz4r\") pod \"openshift-controller-manager-operator-756b6f6bc6-h7kkm\" (UID: \"2720c151-cb93-4622-ad01-379f5f4b0466\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238461 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1878674e-9147-4b42-b95a-6c5a82ff5573-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6bt2z\" (UID: \"1878674e-9147-4b42-b95a-6c5a82ff5573\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238481 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcl6r\" (UniqueName: \"kubernetes.io/projected/6cec3a69-7dbc-4781-a800-452a7e740adc-kube-api-access-mcl6r\") pod \"controller-manager-879f6c89f-b9xb8\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238497 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/68db36cd-270c-4550-8642-44f72796a198-image-import-ca\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238526 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/68db36cd-270c-4550-8642-44f72796a198-node-pullsecrets\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238549 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25rwc\" (UniqueName: \"kubernetes.io/projected/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-kube-api-access-25rwc\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238568 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-l7mfx\" (UID: \"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175\") " pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238585 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238603 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238640 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-config\") pod \"controller-manager-879f6c89f-b9xb8\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238656 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38613c37-a98c-443b-8aaf-c75fcabc9238-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-pmd7x\" (UID: \"38613c37-a98c-443b-8aaf-c75fcabc9238\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238675 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f97c9c68-21cb-4f92-b725-3130f00005c8-etcd-ca\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238694 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f97c9c68-21cb-4f92-b725-3130f00005c8-etcd-service-ca\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238714 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68db36cd-270c-4550-8642-44f72796a198-config\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238733 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2720c151-cb93-4622-ad01-379f5f4b0466-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-h7kkm\" (UID: \"2720c151-cb93-4622-ad01-379f5f4b0466\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238753 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-serving-cert\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238772 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-trusted-ca-bundle\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238791 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238798 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238807 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7jjh\" (UniqueName: \"kubernetes.io/projected/68db36cd-270c-4550-8642-44f72796a198-kube-api-access-z7jjh\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238846 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68d2f0b1-e6a9-4133-aa4d-19c40db638d2-serving-cert\") pod \"openshift-config-operator-7777fb866f-9zvgq\" (UID: \"68d2f0b1-e6a9-4133-aa4d-19c40db638d2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238865 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-oauth-config\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238883 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znxjj\" (UniqueName: \"kubernetes.io/projected/95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a-kube-api-access-znxjj\") pod \"authentication-operator-69f744f599-cs6j4\" (UID: \"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238905 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238922 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c886dc3f-4a59-4e31-a900-2c000427c7a9-config\") pod \"route-controller-manager-6576b87f9c-5pnp5\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238953 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r6cx\" (UniqueName: \"kubernetes.io/projected/c886dc3f-4a59-4e31-a900-2c000427c7a9-kube-api-access-9r6cx\") pod \"route-controller-manager-6576b87f9c-5pnp5\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238969 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkr6g\" (UniqueName: \"kubernetes.io/projected/68d2f0b1-e6a9-4133-aa4d-19c40db638d2-kube-api-access-nkr6g\") pod \"openshift-config-operator-7777fb866f-9zvgq\" (UID: \"68d2f0b1-e6a9-4133-aa4d-19c40db638d2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.238989 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f97c9c68-21cb-4f92-b725-3130f00005c8-config\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.239008 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0234bac5-34c7-4ef3-93f2-4673ebf78d7a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bb5fl\" (UID: \"0234bac5-34c7-4ef3-93f2-4673ebf78d7a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.239027 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm5hq\" (UniqueName: \"kubernetes.io/projected/42bba742-38b4-4d27-9de8-d5f23fe28e84-kube-api-access-cm5hq\") pod \"kube-storage-version-migrator-operator-b67b599dd-85h9w\" (UID: \"42bba742-38b4-4d27-9de8-d5f23fe28e84\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.239044 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-service-ca\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.239061 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.239078 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c886dc3f-4a59-4e31-a900-2c000427c7a9-serving-cert\") pod \"route-controller-manager-6576b87f9c-5pnp5\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.239098 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f97c9c68-21cb-4f92-b725-3130f00005c8-serving-cert\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.239116 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/68db36cd-270c-4550-8642-44f72796a198-audit-dir\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.239134 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e9c01d1-d2a8-4d89-afb9-9599728cb909-config\") pod \"console-operator-58897d9998-s2hr5\" (UID: \"8e9c01d1-d2a8-4d89-afb9-9599728cb909\") " pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.239151 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a-serving-cert\") pod \"authentication-operator-69f744f599-cs6j4\" (UID: \"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.239751 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/93126f7a-dfe3-4864-bc30-ac3629eb2bc2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j47hk\" (UID: \"93126f7a-dfe3-4864-bc30-ac3629eb2bc2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.241452 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.241607 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.242798 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.243594 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.249328 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.249886 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.251033 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.254720 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.257695 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.258485 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.258587 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.258722 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.258790 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.261614 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.262130 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.263743 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-r9gw4"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.264172 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-r9gw4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.269493 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.270035 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.270286 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.270718 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.271533 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fbk7w"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.272113 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fbk7w" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.275340 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.275376 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.275630 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cs6j4"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.276732 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.279897 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.282908 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mjqqv"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.284401 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-w9l4v"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.285831 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-twbgk"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.287315 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l7mfx"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.289610 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.291142 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-2266x"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.291699 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-s2hr5"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.291780 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2266x" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.291807 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.293477 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.295065 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kqk8p"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.296692 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.298615 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-v2m45"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.300806 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t98c7"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.301567 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-s6nss"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.301641 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.302462 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.304015 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.305853 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.306882 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-sqfdh"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.308116 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5dxtl"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.309171 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.310219 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.311836 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-zf9xg"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.313511 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2jcd2"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.314639 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.316031 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.317483 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.319010 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fbk7w"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.320817 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-r9gw4"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.323173 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t98c7"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.324765 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.329072 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.329146 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.331549 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-99fnd"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.334562 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-9r679"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.335195 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-9r679" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.335456 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-99fnd" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.335785 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-99fnd"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.336838 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-9r679"] Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.336993 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.339947 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f97c9c68-21cb-4f92-b725-3130f00005c8-serving-cert\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.339975 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340011 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c886dc3f-4a59-4e31-a900-2c000427c7a9-serving-cert\") pod \"route-controller-manager-6576b87f9c-5pnp5\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340032 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/68db36cd-270c-4550-8642-44f72796a198-audit-dir\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340048 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e9c01d1-d2a8-4d89-afb9-9599728cb909-config\") pod \"console-operator-58897d9998-s2hr5\" (UID: \"8e9c01d1-d2a8-4d89-afb9-9599728cb909\") " pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340063 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a-serving-cert\") pod \"authentication-operator-69f744f599-cs6j4\" (UID: \"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340078 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvvck\" (UniqueName: \"kubernetes.io/projected/29e4371e-f34c-409b-b9db-64de91ddd2ef-kube-api-access-wvvck\") pod \"cluster-image-registry-operator-dc59b4c8b-48dh8\" (UID: \"29e4371e-f34c-409b-b9db-64de91ddd2ef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340093 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/68db36cd-270c-4550-8642-44f72796a198-etcd-client\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340108 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz29k\" (UniqueName: \"kubernetes.io/projected/80a392a8-e94f-4dfa-aefa-3dd05e1051c5-kube-api-access-lz29k\") pod \"cluster-samples-operator-665b6dd947-twbgk\" (UID: \"80a392a8-e94f-4dfa-aefa-3dd05e1051c5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-twbgk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340123 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-oauth-serving-cert\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340138 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e9c01d1-d2a8-4d89-afb9-9599728cb909-serving-cert\") pod \"console-operator-58897d9998-s2hr5\" (UID: \"8e9c01d1-d2a8-4d89-afb9-9599728cb909\") " pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340154 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/80a392a8-e94f-4dfa-aefa-3dd05e1051c5-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-twbgk\" (UID: \"80a392a8-e94f-4dfa-aefa-3dd05e1051c5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-twbgk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340156 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/68db36cd-270c-4550-8642-44f72796a198-audit-dir\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340170 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/217d2158-0c6e-48d9-9713-5499a54655d9-metrics-certs\") pod \"router-default-5444994796-6b8xv\" (UID: \"217d2158-0c6e-48d9-9713-5499a54655d9\") " pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340185 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2bbf\" (UniqueName: \"kubernetes.io/projected/8e9c01d1-d2a8-4d89-afb9-9599728cb909-kube-api-access-d2bbf\") pod \"console-operator-58897d9998-s2hr5\" (UID: \"8e9c01d1-d2a8-4d89-afb9-9599728cb909\") " pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340200 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a-config\") pod \"authentication-operator-69f744f599-cs6j4\" (UID: \"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340215 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42bba742-38b4-4d27-9de8-d5f23fe28e84-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-85h9w\" (UID: \"42bba742-38b4-4d27-9de8-d5f23fe28e84\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340229 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68db36cd-270c-4550-8642-44f72796a198-serving-cert\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340245 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzq7j\" (UniqueName: \"kubernetes.io/projected/217d2158-0c6e-48d9-9713-5499a54655d9-kube-api-access-zzq7j\") pod \"router-default-5444994796-6b8xv\" (UID: \"217d2158-0c6e-48d9-9713-5499a54655d9\") " pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340260 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-l7mfx\" (UID: \"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175\") " pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340279 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6d0c9b6a-66ba-4667-a6ea-833b12934c09-proxy-tls\") pod \"machine-config-operator-74547568cd-8jkjv\" (UID: \"6d0c9b6a-66ba-4667-a6ea-833b12934c09\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340296 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b-proxy-tls\") pod \"machine-config-controller-84d6567774-qvnkw\" (UID: \"7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340312 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-audit-dir\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340327 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/68db36cd-270c-4550-8642-44f72796a198-etcd-serving-ca\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340341 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a-service-ca-bundle\") pod \"authentication-operator-69f744f599-cs6j4\" (UID: \"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340356 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42bba742-38b4-4d27-9de8-d5f23fe28e84-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-85h9w\" (UID: \"42bba742-38b4-4d27-9de8-d5f23fe28e84\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340371 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cs6j4\" (UID: \"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340388 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-audit-policies\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340404 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340434 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-encryption-config\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340450 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cec3a69-7dbc-4781-a800-452a7e740adc-serving-cert\") pod \"controller-manager-879f6c89f-b9xb8\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340466 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-audit-dir\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340481 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vklx7\" (UniqueName: \"kubernetes.io/projected/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-kube-api-access-vklx7\") pod \"marketplace-operator-79b997595-l7mfx\" (UID: \"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175\") " pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340504 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29e4371e-f34c-409b-b9db-64de91ddd2ef-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-48dh8\" (UID: \"29e4371e-f34c-409b-b9db-64de91ddd2ef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340518 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/29e4371e-f34c-409b-b9db-64de91ddd2ef-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-48dh8\" (UID: \"29e4371e-f34c-409b-b9db-64de91ddd2ef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340534 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b-config\") pod \"kube-controller-manager-operator-78b949d7b-ndt7g\" (UID: \"12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340551 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/217d2158-0c6e-48d9-9713-5499a54655d9-default-certificate\") pod \"router-default-5444994796-6b8xv\" (UID: \"217d2158-0c6e-48d9-9713-5499a54655d9\") " pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340569 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/68db36cd-270c-4550-8642-44f72796a198-audit\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340593 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zddvn\" (UniqueName: \"kubernetes.io/projected/6d0c9b6a-66ba-4667-a6ea-833b12934c09-kube-api-access-zddvn\") pod \"machine-config-operator-74547568cd-8jkjv\" (UID: \"6d0c9b6a-66ba-4667-a6ea-833b12934c09\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340612 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqmqh\" (UniqueName: \"kubernetes.io/projected/1878674e-9147-4b42-b95a-6c5a82ff5573-kube-api-access-nqmqh\") pod \"openshift-apiserver-operator-796bbdcf4f-6bt2z\" (UID: \"1878674e-9147-4b42-b95a-6c5a82ff5573\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340628 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/68db36cd-270c-4550-8642-44f72796a198-encryption-config\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340645 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-audit-policies\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340663 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340718 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtmhz\" (UniqueName: \"kubernetes.io/projected/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-kube-api-access-wtmhz\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340737 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f97c9c68-21cb-4f92-b725-3130f00005c8-etcd-client\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340753 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0234bac5-34c7-4ef3-93f2-4673ebf78d7a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bb5fl\" (UID: \"0234bac5-34c7-4ef3-93f2-4673ebf78d7a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340768 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/351ebc2b-c707-4aec-8f6b-e79af328fdf5-config\") pod \"machine-approver-56656f9798-mnwj9\" (UID: \"351ebc2b-c707-4aec-8f6b-e79af328fdf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340786 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340806 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t66tc\" (UniqueName: \"kubernetes.io/projected/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-kube-api-access-t66tc\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340810 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-audit-dir\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340824 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c886dc3f-4a59-4e31-a900-2c000427c7a9-client-ca\") pod \"route-controller-manager-6576b87f9c-5pnp5\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340874 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340890 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-audit-dir\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340918 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-client-ca\") pod \"controller-manager-879f6c89f-b9xb8\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.340998 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/351ebc2b-c707-4aec-8f6b-e79af328fdf5-machine-approver-tls\") pod \"machine-approver-56656f9798-mnwj9\" (UID: \"351ebc2b-c707-4aec-8f6b-e79af328fdf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341038 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/68d2f0b1-e6a9-4133-aa4d-19c40db638d2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9zvgq\" (UID: \"68d2f0b1-e6a9-4133-aa4d-19c40db638d2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341077 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/29e4371e-f34c-409b-b9db-64de91ddd2ef-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-48dh8\" (UID: \"29e4371e-f34c-409b-b9db-64de91ddd2ef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341115 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-ndt7g\" (UID: \"12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341149 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-ndt7g\" (UID: \"12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341181 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/68db36cd-270c-4550-8642-44f72796a198-trusted-ca-bundle\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341221 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341255 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4kd7\" (UniqueName: \"kubernetes.io/projected/b2995c47-f228-411e-857d-c2830c577ffc-kube-api-access-g4kd7\") pod \"dns-operator-744455d44c-sqfdh\" (UID: \"b2995c47-f228-411e-857d-c2830c577ffc\") " pod="openshift-dns-operator/dns-operator-744455d44c-sqfdh" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341286 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-etcd-client\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341320 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/217d2158-0c6e-48d9-9713-5499a54655d9-stats-auth\") pod \"router-default-5444994796-6b8xv\" (UID: \"217d2158-0c6e-48d9-9713-5499a54655d9\") " pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341367 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-b9xb8\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341457 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlnxt\" (UniqueName: \"kubernetes.io/projected/7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b-kube-api-access-dlnxt\") pod \"machine-config-controller-84d6567774-qvnkw\" (UID: \"7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341507 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6d0c9b6a-66ba-4667-a6ea-833b12934c09-images\") pod \"machine-config-operator-74547568cd-8jkjv\" (UID: \"6d0c9b6a-66ba-4667-a6ea-833b12934c09\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341537 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/68db36cd-270c-4550-8642-44f72796a198-etcd-serving-ca\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341554 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/351ebc2b-c707-4aec-8f6b-e79af328fdf5-auth-proxy-config\") pod \"machine-approver-56656f9798-mnwj9\" (UID: \"351ebc2b-c707-4aec-8f6b-e79af328fdf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341602 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qvnkw\" (UID: \"7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341632 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-oauth-serving-cert\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341651 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2ljj\" (UniqueName: \"kubernetes.io/projected/f97c9c68-21cb-4f92-b725-3130f00005c8-kube-api-access-d2ljj\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341691 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8e9c01d1-d2a8-4d89-afb9-9599728cb909-trusted-ca\") pod \"console-operator-58897d9998-s2hr5\" (UID: \"8e9c01d1-d2a8-4d89-afb9-9599728cb909\") " pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341733 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341787 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341823 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341870 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b2995c47-f228-411e-857d-c2830c577ffc-metrics-tls\") pod \"dns-operator-744455d44c-sqfdh\" (UID: \"b2995c47-f228-411e-857d-c2830c577ffc\") " pod="openshift-dns-operator/dns-operator-744455d44c-sqfdh" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341926 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/38613c37-a98c-443b-8aaf-c75fcabc9238-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-pmd7x\" (UID: \"38613c37-a98c-443b-8aaf-c75fcabc9238\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341960 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-config\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.341993 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/217d2158-0c6e-48d9-9713-5499a54655d9-service-ca-bundle\") pod \"router-default-5444994796-6b8xv\" (UID: \"217d2158-0c6e-48d9-9713-5499a54655d9\") " pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342030 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6d0c9b6a-66ba-4667-a6ea-833b12934c09-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8jkjv\" (UID: \"6d0c9b6a-66ba-4667-a6ea-833b12934c09\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342063 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2720c151-cb93-4622-ad01-379f5f4b0466-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-h7kkm\" (UID: \"2720c151-cb93-4622-ad01-379f5f4b0466\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342096 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0234bac5-34c7-4ef3-93f2-4673ebf78d7a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bb5fl\" (UID: \"0234bac5-34c7-4ef3-93f2-4673ebf78d7a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342128 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38613c37-a98c-443b-8aaf-c75fcabc9238-config\") pod \"kube-apiserver-operator-766d6c64bb-pmd7x\" (UID: \"38613c37-a98c-443b-8aaf-c75fcabc9238\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342160 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48x5l\" (UniqueName: \"kubernetes.io/projected/351ebc2b-c707-4aec-8f6b-e79af328fdf5-kube-api-access-48x5l\") pod \"machine-approver-56656f9798-mnwj9\" (UID: \"351ebc2b-c707-4aec-8f6b-e79af328fdf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342193 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-serving-cert\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342227 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1878674e-9147-4b42-b95a-6c5a82ff5573-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6bt2z\" (UID: \"1878674e-9147-4b42-b95a-6c5a82ff5573\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342262 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk5xf\" (UniqueName: \"kubernetes.io/projected/67608617-45fd-4c45-87f9-325996855e5e-kube-api-access-lk5xf\") pod \"downloads-7954f5f757-s6nss\" (UID: \"67608617-45fd-4c45-87f9-325996855e5e\") " pod="openshift-console/downloads-7954f5f757-s6nss" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342294 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1878674e-9147-4b42-b95a-6c5a82ff5573-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6bt2z\" (UID: \"1878674e-9147-4b42-b95a-6c5a82ff5573\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342332 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcl6r\" (UniqueName: \"kubernetes.io/projected/6cec3a69-7dbc-4781-a800-452a7e740adc-kube-api-access-mcl6r\") pod \"controller-manager-879f6c89f-b9xb8\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342367 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtz4r\" (UniqueName: \"kubernetes.io/projected/2720c151-cb93-4622-ad01-379f5f4b0466-kube-api-access-mtz4r\") pod \"openshift-controller-manager-operator-756b6f6bc6-h7kkm\" (UID: \"2720c151-cb93-4622-ad01-379f5f4b0466\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342399 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/68db36cd-270c-4550-8642-44f72796a198-image-import-ca\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342463 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/68db36cd-270c-4550-8642-44f72796a198-node-pullsecrets\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342502 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-l7mfx\" (UID: \"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175\") " pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342535 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25rwc\" (UniqueName: \"kubernetes.io/projected/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-kube-api-access-25rwc\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342570 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342602 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342612 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e9c01d1-d2a8-4d89-afb9-9599728cb909-config\") pod \"console-operator-58897d9998-s2hr5\" (UID: \"8e9c01d1-d2a8-4d89-afb9-9599728cb909\") " pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342663 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38613c37-a98c-443b-8aaf-c75fcabc9238-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-pmd7x\" (UID: \"38613c37-a98c-443b-8aaf-c75fcabc9238\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342687 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-config\") pod \"controller-manager-879f6c89f-b9xb8\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342715 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f97c9c68-21cb-4f92-b725-3130f00005c8-etcd-ca\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342739 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f97c9c68-21cb-4f92-b725-3130f00005c8-etcd-service-ca\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342761 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68db36cd-270c-4550-8642-44f72796a198-config\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342787 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-serving-cert\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342810 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2720c151-cb93-4622-ad01-379f5f4b0466-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-h7kkm\" (UID: \"2720c151-cb93-4622-ad01-379f5f4b0466\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342838 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-trusted-ca-bundle\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342860 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342863 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f97c9c68-21cb-4f92-b725-3130f00005c8-serving-cert\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342886 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7jjh\" (UniqueName: \"kubernetes.io/projected/68db36cd-270c-4550-8642-44f72796a198-kube-api-access-z7jjh\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342911 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68d2f0b1-e6a9-4133-aa4d-19c40db638d2-serving-cert\") pod \"openshift-config-operator-7777fb866f-9zvgq\" (UID: \"68d2f0b1-e6a9-4133-aa4d-19c40db638d2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342936 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-oauth-config\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342959 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znxjj\" (UniqueName: \"kubernetes.io/projected/95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a-kube-api-access-znxjj\") pod \"authentication-operator-69f744f599-cs6j4\" (UID: \"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.343002 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.343033 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c886dc3f-4a59-4e31-a900-2c000427c7a9-config\") pod \"route-controller-manager-6576b87f9c-5pnp5\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.343055 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkr6g\" (UniqueName: \"kubernetes.io/projected/68d2f0b1-e6a9-4133-aa4d-19c40db638d2-kube-api-access-nkr6g\") pod \"openshift-config-operator-7777fb866f-9zvgq\" (UID: \"68d2f0b1-e6a9-4133-aa4d-19c40db638d2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.343076 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r6cx\" (UniqueName: \"kubernetes.io/projected/c886dc3f-4a59-4e31-a900-2c000427c7a9-kube-api-access-9r6cx\") pod \"route-controller-manager-6576b87f9c-5pnp5\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.343094 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f97c9c68-21cb-4f92-b725-3130f00005c8-config\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.343113 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0234bac5-34c7-4ef3-93f2-4673ebf78d7a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bb5fl\" (UID: \"0234bac5-34c7-4ef3-93f2-4673ebf78d7a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.343128 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-service-ca\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.343146 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm5hq\" (UniqueName: \"kubernetes.io/projected/42bba742-38b4-4d27-9de8-d5f23fe28e84-kube-api-access-cm5hq\") pod \"kube-storage-version-migrator-operator-b67b599dd-85h9w\" (UID: \"42bba742-38b4-4d27-9de8-d5f23fe28e84\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.343170 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a-serving-cert\") pod \"authentication-operator-69f744f599-cs6j4\" (UID: \"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.343342 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.343819 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a-config\") pod \"authentication-operator-69f744f599-cs6j4\" (UID: \"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.343897 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/80a392a8-e94f-4dfa-aefa-3dd05e1051c5-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-twbgk\" (UID: \"80a392a8-e94f-4dfa-aefa-3dd05e1051c5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-twbgk" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.344016 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/217d2158-0c6e-48d9-9713-5499a54655d9-metrics-certs\") pod \"router-default-5444994796-6b8xv\" (UID: \"217d2158-0c6e-48d9-9713-5499a54655d9\") " pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.344834 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.344894 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cs6j4\" (UID: \"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.345056 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-config\") pod \"controller-manager-879f6c89f-b9xb8\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.345211 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29e4371e-f34c-409b-b9db-64de91ddd2ef-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-48dh8\" (UID: \"29e4371e-f34c-409b-b9db-64de91ddd2ef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.345363 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-audit-policies\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.345606 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f97c9c68-21cb-4f92-b725-3130f00005c8-etcd-ca\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.345602 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.345737 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/68db36cd-270c-4550-8642-44f72796a198-audit\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.345841 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.346063 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68db36cd-270c-4550-8642-44f72796a198-serving-cert\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.346106 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f97c9c68-21cb-4f92-b725-3130f00005c8-etcd-service-ca\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.346777 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/68db36cd-270c-4550-8642-44f72796a198-etcd-client\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.346965 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68db36cd-270c-4550-8642-44f72796a198-config\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.347179 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/68db36cd-270c-4550-8642-44f72796a198-node-pullsecrets\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.347240 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-audit-policies\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.348247 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-config\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.348283 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c886dc3f-4a59-4e31-a900-2c000427c7a9-config\") pod \"route-controller-manager-6576b87f9c-5pnp5\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.348453 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/68db36cd-270c-4550-8642-44f72796a198-image-import-ca\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.348588 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c886dc3f-4a59-4e31-a900-2c000427c7a9-client-ca\") pod \"route-controller-manager-6576b87f9c-5pnp5\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.348769 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-trusted-ca-bundle\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.348901 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.348984 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/217d2158-0c6e-48d9-9713-5499a54655d9-service-ca-bundle\") pod \"router-default-5444994796-6b8xv\" (UID: \"217d2158-0c6e-48d9-9713-5499a54655d9\") " pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.349176 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f97c9c68-21cb-4f92-b725-3130f00005c8-config\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.349484 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-client-ca\") pod \"controller-manager-879f6c89f-b9xb8\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.349496 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.349705 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-service-ca\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.349786 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.349837 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6d0c9b6a-66ba-4667-a6ea-833b12934c09-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8jkjv\" (UID: \"6d0c9b6a-66ba-4667-a6ea-833b12934c09\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.342096 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a-service-ca-bundle\") pod \"authentication-operator-69f744f599-cs6j4\" (UID: \"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.350020 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2720c151-cb93-4622-ad01-379f5f4b0466-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-h7kkm\" (UID: \"2720c151-cb93-4622-ad01-379f5f4b0466\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.350137 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/68d2f0b1-e6a9-4133-aa4d-19c40db638d2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9zvgq\" (UID: \"68d2f0b1-e6a9-4133-aa4d-19c40db638d2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.350205 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/217d2158-0c6e-48d9-9713-5499a54655d9-default-certificate\") pod \"router-default-5444994796-6b8xv\" (UID: \"217d2158-0c6e-48d9-9713-5499a54655d9\") " pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.350984 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qvnkw\" (UID: \"7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.351075 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2720c151-cb93-4622-ad01-379f5f4b0466-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-h7kkm\" (UID: \"2720c151-cb93-4622-ad01-379f5f4b0466\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.351139 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/68db36cd-270c-4550-8642-44f72796a198-trusted-ca-bundle\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.351272 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/68db36cd-270c-4550-8642-44f72796a198-encryption-config\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.351304 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1878674e-9147-4b42-b95a-6c5a82ff5573-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6bt2z\" (UID: \"1878674e-9147-4b42-b95a-6c5a82ff5573\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.351595 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8e9c01d1-d2a8-4d89-afb9-9599728cb909-trusted-ca\") pod \"console-operator-58897d9998-s2hr5\" (UID: \"8e9c01d1-d2a8-4d89-afb9-9599728cb909\") " pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.351761 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.352106 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1878674e-9147-4b42-b95a-6c5a82ff5573-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6bt2z\" (UID: \"1878674e-9147-4b42-b95a-6c5a82ff5573\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.352260 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-b9xb8\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.352315 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-etcd-client\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.352390 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f97c9c68-21cb-4f92-b725-3130f00005c8-etcd-client\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.353244 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68d2f0b1-e6a9-4133-aa4d-19c40db638d2-serving-cert\") pod \"openshift-config-operator-7777fb866f-9zvgq\" (UID: \"68d2f0b1-e6a9-4133-aa4d-19c40db638d2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.353557 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.353598 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-oauth-config\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.353718 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e9c01d1-d2a8-4d89-afb9-9599728cb909-serving-cert\") pod \"console-operator-58897d9998-s2hr5\" (UID: \"8e9c01d1-d2a8-4d89-afb9-9599728cb909\") " pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.353845 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/29e4371e-f34c-409b-b9db-64de91ddd2ef-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-48dh8\" (UID: \"29e4371e-f34c-409b-b9db-64de91ddd2ef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.354016 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-serving-cert\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.354299 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.354780 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/217d2158-0c6e-48d9-9713-5499a54655d9-stats-auth\") pod \"router-default-5444994796-6b8xv\" (UID: \"217d2158-0c6e-48d9-9713-5499a54655d9\") " pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.355290 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c886dc3f-4a59-4e31-a900-2c000427c7a9-serving-cert\") pod \"route-controller-manager-6576b87f9c-5pnp5\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.355525 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-encryption-config\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.356025 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.356157 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.356847 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.357138 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cec3a69-7dbc-4781-a800-452a7e740adc-serving-cert\") pod \"controller-manager-879f6c89f-b9xb8\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.357803 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-serving-cert\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.358321 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.375662 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.396116 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.400079 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0234bac5-34c7-4ef3-93f2-4673ebf78d7a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bb5fl\" (UID: \"0234bac5-34c7-4ef3-93f2-4673ebf78d7a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.415808 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.420146 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0234bac5-34c7-4ef3-93f2-4673ebf78d7a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bb5fl\" (UID: \"0234bac5-34c7-4ef3-93f2-4673ebf78d7a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.435403 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.455964 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.460411 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/38613c37-a98c-443b-8aaf-c75fcabc9238-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-pmd7x\" (UID: \"38613c37-a98c-443b-8aaf-c75fcabc9238\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.475636 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.476054 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.496406 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.500745 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38613c37-a98c-443b-8aaf-c75fcabc9238-config\") pod \"kube-apiserver-operator-766d6c64bb-pmd7x\" (UID: \"38613c37-a98c-443b-8aaf-c75fcabc9238\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.516756 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.536271 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.542856 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b2995c47-f228-411e-857d-c2830c577ffc-metrics-tls\") pod \"dns-operator-744455d44c-sqfdh\" (UID: \"b2995c47-f228-411e-857d-c2830c577ffc\") " pod="openshift-dns-operator/dns-operator-744455d44c-sqfdh" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.556288 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.576138 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.595994 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.616685 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.636661 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.645572 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-ndt7g\" (UID: \"12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.656069 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.675823 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.686397 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b-config\") pod \"kube-controller-manager-operator-78b949d7b-ndt7g\" (UID: \"12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.695969 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.705536 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b-proxy-tls\") pod \"machine-config-controller-84d6567774-qvnkw\" (UID: \"7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.716027 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.735547 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.756008 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.776727 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.781383 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6d0c9b6a-66ba-4667-a6ea-833b12934c09-images\") pod \"machine-config-operator-74547568cd-8jkjv\" (UID: \"6d0c9b6a-66ba-4667-a6ea-833b12934c09\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.796589 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.815632 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.823927 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6d0c9b6a-66ba-4667-a6ea-833b12934c09-proxy-tls\") pod \"machine-config-operator-74547568cd-8jkjv\" (UID: \"6d0c9b6a-66ba-4667-a6ea-833b12934c09\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.846866 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.849817 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-l7mfx\" (UID: \"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175\") " pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.857231 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.876593 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.888254 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-l7mfx\" (UID: \"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175\") " pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.895897 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.917883 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.935961 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.942322 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/351ebc2b-c707-4aec-8f6b-e79af328fdf5-auth-proxy-config\") pod \"machine-approver-56656f9798-mnwj9\" (UID: \"351ebc2b-c707-4aec-8f6b-e79af328fdf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.956488 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.976329 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.984641 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/351ebc2b-c707-4aec-8f6b-e79af328fdf5-machine-approver-tls\") pod \"machine-approver-56656f9798-mnwj9\" (UID: \"351ebc2b-c707-4aec-8f6b-e79af328fdf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.997285 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 15:36:53 crc kubenswrapper[4739]: I1201 15:36:53.998228 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/351ebc2b-c707-4aec-8f6b-e79af328fdf5-config\") pod \"machine-approver-56656f9798-mnwj9\" (UID: \"351ebc2b-c707-4aec-8f6b-e79af328fdf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.016792 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.036889 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.057145 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.065128 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42bba742-38b4-4d27-9de8-d5f23fe28e84-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-85h9w\" (UID: \"42bba742-38b4-4d27-9de8-d5f23fe28e84\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.077225 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.097046 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.106713 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42bba742-38b4-4d27-9de8-d5f23fe28e84-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-85h9w\" (UID: \"42bba742-38b4-4d27-9de8-d5f23fe28e84\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.117072 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.136798 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.177342 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.194701 4739 request.go:700] Waited for 1.013551442s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/secrets?fieldSelector=metadata.name%3Dcontrol-plane-machine-set-operator-dockercfg-k9rxt&limit=500&resourceVersion=0 Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.196632 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.216462 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.236409 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.297347 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.304267 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpdjx\" (UniqueName: \"kubernetes.io/projected/93126f7a-dfe3-4864-bc30-ac3629eb2bc2-kube-api-access-vpdjx\") pod \"machine-api-operator-5694c8668f-j47hk\" (UID: \"93126f7a-dfe3-4864-bc30-ac3629eb2bc2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.325133 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.336761 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.356899 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.376526 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.396401 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.416995 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.437016 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.457019 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.476532 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.476624 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.476530 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.477191 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.496772 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.516908 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.537352 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.557086 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.576874 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.585847 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.597469 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.617048 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.637097 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.655966 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.677178 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.697397 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.718684 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.736796 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.757038 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.776286 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.797813 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.817244 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.819085 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j47hk"] Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.835665 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.855842 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.877648 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.896168 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.916873 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.936085 4739 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.957847 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.976660 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 15:36:54 crc kubenswrapper[4739]: I1201 15:36:54.996521 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.017506 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.036953 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.056334 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.075806 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.095749 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.095928 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" event={"ID":"93126f7a-dfe3-4864-bc30-ac3629eb2bc2","Type":"ContainerStarted","Data":"4f93fb7e5ccadb0262b9b0c1149e70fafa715a5ca5aa981507a03a839e0bde68"} Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.096000 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" event={"ID":"93126f7a-dfe3-4864-bc30-ac3629eb2bc2","Type":"ContainerStarted","Data":"73756c21c1727e81e46b2b44d727a18655fd80f99e129013a958ded945dfc425"} Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.096016 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" event={"ID":"93126f7a-dfe3-4864-bc30-ac3629eb2bc2","Type":"ContainerStarted","Data":"5d20287681c15cb08b40aabd54b80dd2c0e69ecf807028c90f3ce08e13195635"} Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.133595 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz29k\" (UniqueName: \"kubernetes.io/projected/80a392a8-e94f-4dfa-aefa-3dd05e1051c5-kube-api-access-lz29k\") pod \"cluster-samples-operator-665b6dd947-twbgk\" (UID: \"80a392a8-e94f-4dfa-aefa-3dd05e1051c5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-twbgk" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.152798 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvvck\" (UniqueName: \"kubernetes.io/projected/29e4371e-f34c-409b-b9db-64de91ddd2ef-kube-api-access-wvvck\") pod \"cluster-image-registry-operator-dc59b4c8b-48dh8\" (UID: \"29e4371e-f34c-409b-b9db-64de91ddd2ef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.173973 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm5hq\" (UniqueName: \"kubernetes.io/projected/42bba742-38b4-4d27-9de8-d5f23fe28e84-kube-api-access-cm5hq\") pod \"kube-storage-version-migrator-operator-b67b599dd-85h9w\" (UID: \"42bba742-38b4-4d27-9de8-d5f23fe28e84\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.193923 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38613c37-a98c-443b-8aaf-c75fcabc9238-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-pmd7x\" (UID: \"38613c37-a98c-443b-8aaf-c75fcabc9238\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.209953 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.215213 4739 request.go:700] Waited for 1.871009834s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/serviceaccounts/router/token Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.223120 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2bbf\" (UniqueName: \"kubernetes.io/projected/8e9c01d1-d2a8-4d89-afb9-9599728cb909-kube-api-access-d2bbf\") pod \"console-operator-58897d9998-s2hr5\" (UID: \"8e9c01d1-d2a8-4d89-afb9-9599728cb909\") " pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.245406 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzq7j\" (UniqueName: \"kubernetes.io/projected/217d2158-0c6e-48d9-9713-5499a54655d9-kube-api-access-zzq7j\") pod \"router-default-5444994796-6b8xv\" (UID: \"217d2158-0c6e-48d9-9713-5499a54655d9\") " pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.254639 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vklx7\" (UniqueName: \"kubernetes.io/projected/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-kube-api-access-vklx7\") pod \"marketplace-operator-79b997595-l7mfx\" (UID: \"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175\") " pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.272713 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.283241 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48x5l\" (UniqueName: \"kubernetes.io/projected/351ebc2b-c707-4aec-8f6b-e79af328fdf5-kube-api-access-48x5l\") pod \"machine-approver-56656f9798-mnwj9\" (UID: \"351ebc2b-c707-4aec-8f6b-e79af328fdf5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.294196 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/29e4371e-f34c-409b-b9db-64de91ddd2ef-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-48dh8\" (UID: \"29e4371e-f34c-409b-b9db-64de91ddd2ef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.302605 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-twbgk" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.320692 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znxjj\" (UniqueName: \"kubernetes.io/projected/95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a-kube-api-access-znxjj\") pod \"authentication-operator-69f744f599-cs6j4\" (UID: \"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.333175 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtz4r\" (UniqueName: \"kubernetes.io/projected/2720c151-cb93-4622-ad01-379f5f4b0466-kube-api-access-mtz4r\") pod \"openshift-controller-manager-operator-756b6f6bc6-h7kkm\" (UID: \"2720c151-cb93-4622-ad01-379f5f4b0466\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.358781 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zddvn\" (UniqueName: \"kubernetes.io/projected/6d0c9b6a-66ba-4667-a6ea-833b12934c09-kube-api-access-zddvn\") pod \"machine-config-operator-74547568cd-8jkjv\" (UID: \"6d0c9b6a-66ba-4667-a6ea-833b12934c09\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.371146 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqmqh\" (UniqueName: \"kubernetes.io/projected/1878674e-9147-4b42-b95a-6c5a82ff5573-kube-api-access-nqmqh\") pod \"openshift-apiserver-operator-796bbdcf4f-6bt2z\" (UID: \"1878674e-9147-4b42-b95a-6c5a82ff5573\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.373589 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.391107 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk5xf\" (UniqueName: \"kubernetes.io/projected/67608617-45fd-4c45-87f9-325996855e5e-kube-api-access-lk5xf\") pod \"downloads-7954f5f757-s6nss\" (UID: \"67608617-45fd-4c45-87f9-325996855e5e\") " pod="openshift-console/downloads-7954f5f757-s6nss" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.392208 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.400936 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-s6nss" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.412070 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25rwc\" (UniqueName: \"kubernetes.io/projected/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-kube-api-access-25rwc\") pod \"oauth-openshift-558db77b4-mjqqv\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.425208 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.431803 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4kd7\" (UniqueName: \"kubernetes.io/projected/b2995c47-f228-411e-857d-c2830c577ffc-kube-api-access-g4kd7\") pod \"dns-operator-744455d44c-sqfdh\" (UID: \"b2995c47-f228-411e-857d-c2830c577ffc\") " pod="openshift-dns-operator/dns-operator-744455d44c-sqfdh" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.435305 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.455271 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t66tc\" (UniqueName: \"kubernetes.io/projected/8d2be141-5a5f-4bde-b56f-83ea18ff8e99-kube-api-access-t66tc\") pod \"apiserver-7bbb656c7d-8bq5d\" (UID: \"8d2be141-5a5f-4bde-b56f-83ea18ff8e99\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.462034 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x"] Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.470128 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtmhz\" (UniqueName: \"kubernetes.io/projected/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-kube-api-access-wtmhz\") pod \"console-f9d7485db-5dxtl\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.491738 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.494953 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkr6g\" (UniqueName: \"kubernetes.io/projected/68d2f0b1-e6a9-4133-aa4d-19c40db638d2-kube-api-access-nkr6g\") pod \"openshift-config-operator-7777fb866f-9zvgq\" (UID: \"68d2f0b1-e6a9-4133-aa4d-19c40db638d2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.499891 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.509989 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.512704 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r6cx\" (UniqueName: \"kubernetes.io/projected/c886dc3f-4a59-4e31-a900-2c000427c7a9-kube-api-access-9r6cx\") pod \"route-controller-manager-6576b87f9c-5pnp5\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.518208 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-sqfdh" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.539738 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.540705 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0234bac5-34c7-4ef3-93f2-4673ebf78d7a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bb5fl\" (UID: \"0234bac5-34c7-4ef3-93f2-4673ebf78d7a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.549696 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.562660 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.565616 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7jjh\" (UniqueName: \"kubernetes.io/projected/68db36cd-270c-4550-8642-44f72796a198-kube-api-access-z7jjh\") pod \"apiserver-76f77b778f-w9l4v\" (UID: \"68db36cd-270c-4550-8642-44f72796a198\") " pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.571625 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcl6r\" (UniqueName: \"kubernetes.io/projected/6cec3a69-7dbc-4781-a800-452a7e740adc-kube-api-access-mcl6r\") pod \"controller-manager-879f6c89f-b9xb8\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.577952 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w"] Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.581151 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.595639 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2ljj\" (UniqueName: \"kubernetes.io/projected/f97c9c68-21cb-4f92-b725-3130f00005c8-kube-api-access-d2ljj\") pod \"etcd-operator-b45778765-zf9xg\" (UID: \"f97c9c68-21cb-4f92-b725-3130f00005c8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.614407 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-ndt7g\" (UID: \"12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.621060 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.636450 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.642911 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-twbgk"] Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.652122 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.654103 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlnxt\" (UniqueName: \"kubernetes.io/projected/7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b-kube-api-access-dlnxt\") pod \"machine-config-controller-84d6567774-qvnkw\" (UID: \"7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.655668 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.662616 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.664021 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cs6j4"] Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.682143 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.693811 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/54c65dfa-5787-4f1f-979c-9e5daf517650-registry-certificates\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.693844 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-registry-tls\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.693868 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/54c65dfa-5787-4f1f-979c-9e5daf517650-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.693890 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26wk4\" (UniqueName: \"kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-kube-api-access-26wk4\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.693930 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/54c65dfa-5787-4f1f-979c-9e5daf517650-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.693959 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54c65dfa-5787-4f1f-979c-9e5daf517650-trusted-ca\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.693974 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-bound-sa-token\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.694012 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: E1201 15:36:55.694259 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:56.194248557 +0000 UTC m=+118.019994651 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.709026 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.716771 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.717164 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.736646 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.756562 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.778204 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.794937 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:55 crc kubenswrapper[4739]: E1201 15:36:55.795058 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:56.295039966 +0000 UTC m=+118.120786060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795126 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sjf6\" (UniqueName: \"kubernetes.io/projected/347982ab-7ed7-4b00-ae43-03b9b91b5133-kube-api-access-7sjf6\") pod \"olm-operator-6b444d44fb-f55t6\" (UID: \"347982ab-7ed7-4b00-ae43-03b9b91b5133\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795185 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djr8g\" (UniqueName: \"kubernetes.io/projected/8a984dbf-3bbf-475a-9c5e-a7b98f594eaf-kube-api-access-djr8g\") pod \"packageserver-d55dfcdfc-8p4jr\" (UID: \"8a984dbf-3bbf-475a-9c5e-a7b98f594eaf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795216 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f7e22852-525d-453f-a329-c15a1c5977ad-config-volume\") pod \"collect-profiles-29410050-688rv\" (UID: \"f7e22852-525d-453f-a329-c15a1c5977ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795233 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/19abbd38-da74-4e94-94c5-9461701ac88d-socket-dir\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795247 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/19abbd38-da74-4e94-94c5-9461701ac88d-registration-dir\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795274 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4s66\" (UniqueName: \"kubernetes.io/projected/c5277ea8-3db8-406c-9715-d6ce69e75470-kube-api-access-w4s66\") pod \"migrator-59844c95c7-fbk7w\" (UID: \"c5277ea8-3db8-406c-9715-d6ce69e75470\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fbk7w" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795301 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/347982ab-7ed7-4b00-ae43-03b9b91b5133-profile-collector-cert\") pod \"olm-operator-6b444d44fb-f55t6\" (UID: \"347982ab-7ed7-4b00-ae43-03b9b91b5133\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795317 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b448b273-61d9-4418-b752-30f6e6556217-certs\") pod \"machine-config-server-2266x\" (UID: \"b448b273-61d9-4418-b752-30f6e6556217\") " pod="openshift-machine-config-operator/machine-config-server-2266x" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795346 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/54c65dfa-5787-4f1f-979c-9e5daf517650-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795462 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ec35359a-8d42-4230-8f34-1dcbee21c8f3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-v2m45\" (UID: \"ec35359a-8d42-4230-8f34-1dcbee21c8f3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-v2m45" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795481 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h6j5\" (UniqueName: \"kubernetes.io/projected/a90528ef-02f4-4012-bff1-1da159712d79-kube-api-access-4h6j5\") pod \"service-ca-operator-777779d784-bgbt2\" (UID: \"a90528ef-02f4-4012-bff1-1da159712d79\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795519 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh9ch\" (UniqueName: \"kubernetes.io/projected/ec35359a-8d42-4230-8f34-1dcbee21c8f3-kube-api-access-mh9ch\") pod \"multus-admission-controller-857f4d67dd-v2m45\" (UID: \"ec35359a-8d42-4230-8f34-1dcbee21c8f3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-v2m45" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795588 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54c65dfa-5787-4f1f-979c-9e5daf517650-trusted-ca\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795604 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-bound-sa-token\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795631 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/19abbd38-da74-4e94-94c5-9461701ac88d-mountpoint-dir\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795665 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tf5k\" (UniqueName: \"kubernetes.io/projected/91f97449-8808-4e97-980b-3452f9cb1983-kube-api-access-2tf5k\") pod \"service-ca-9c57cc56f-r9gw4\" (UID: \"91f97449-8808-4e97-980b-3452f9cb1983\") " pod="openshift-service-ca/service-ca-9c57cc56f-r9gw4" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795705 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795723 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-775d9\" (UniqueName: \"kubernetes.io/projected/8794d67a-f5f8-4f02-a5d4-8b1f214355bc-kube-api-access-775d9\") pod \"catalog-operator-68c6474976-l2r9n\" (UID: \"8794d67a-f5f8-4f02-a5d4-8b1f214355bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795825 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/91f97449-8808-4e97-980b-3452f9cb1983-signing-key\") pod \"service-ca-9c57cc56f-r9gw4\" (UID: \"91f97449-8808-4e97-980b-3452f9cb1983\") " pod="openshift-service-ca/service-ca-9c57cc56f-r9gw4" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795850 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwdcg\" (UniqueName: \"kubernetes.io/projected/f7e22852-525d-453f-a329-c15a1c5977ad-kube-api-access-kwdcg\") pod \"collect-profiles-29410050-688rv\" (UID: \"f7e22852-525d-453f-a329-c15a1c5977ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795893 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a03bbd9-1c8c-4382-ad9d-c59e2efc9427-config-volume\") pod \"dns-default-99fnd\" (UID: \"8a03bbd9-1c8c-4382-ad9d-c59e2efc9427\") " pod="openshift-dns/dns-default-99fnd" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795943 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpwln\" (UniqueName: \"kubernetes.io/projected/80167c35-d69a-44bc-8757-87b7ebf137ac-kube-api-access-dpwln\") pod \"ingress-operator-5b745b69d9-2vr2v\" (UID: \"80167c35-d69a-44bc-8757-87b7ebf137ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795958 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/19abbd38-da74-4e94-94c5-9461701ac88d-csi-data-dir\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.795999 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8a984dbf-3bbf-475a-9c5e-a7b98f594eaf-webhook-cert\") pod \"packageserver-d55dfcdfc-8p4jr\" (UID: \"8a984dbf-3bbf-475a-9c5e-a7b98f594eaf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.796042 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8a03bbd9-1c8c-4382-ad9d-c59e2efc9427-metrics-tls\") pod \"dns-default-99fnd\" (UID: \"8a03bbd9-1c8c-4382-ad9d-c59e2efc9427\") " pod="openshift-dns/dns-default-99fnd" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.797022 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f7e22852-525d-453f-a329-c15a1c5977ad-secret-volume\") pod \"collect-profiles-29410050-688rv\" (UID: \"f7e22852-525d-453f-a329-c15a1c5977ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.797147 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwr2q\" (UniqueName: \"kubernetes.io/projected/c234ec14-adc4-4089-b41f-0df1e4010599-kube-api-access-fwr2q\") pod \"ingress-canary-9r679\" (UID: \"c234ec14-adc4-4089-b41f-0df1e4010599\") " pod="openshift-ingress-canary/ingress-canary-9r679" Dec 01 15:36:55 crc kubenswrapper[4739]: E1201 15:36:55.800537 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:56.300519529 +0000 UTC m=+118.126265623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.802710 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.804398 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54c65dfa-5787-4f1f-979c-9e5daf517650-trusted-ca\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.805466 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b448b273-61d9-4418-b752-30f6e6556217-node-bootstrap-token\") pod \"machine-config-server-2266x\" (UID: \"b448b273-61d9-4418-b752-30f6e6556217\") " pod="openshift-machine-config-operator/machine-config-server-2266x" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.813294 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xrqs\" (UniqueName: \"kubernetes.io/projected/c0af2f6a-1320-4a72-9bfc-684d0f666756-kube-api-access-6xrqs\") pod \"control-plane-machine-set-operator-78cbb6b69f-2jcd2\" (UID: \"c0af2f6a-1320-4a72-9bfc-684d0f666756\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2jcd2" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.813371 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/54c65dfa-5787-4f1f-979c-9e5daf517650-registry-certificates\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.813396 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/80167c35-d69a-44bc-8757-87b7ebf137ac-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2vr2v\" (UID: \"80167c35-d69a-44bc-8757-87b7ebf137ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.813599 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-registry-tls\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.813628 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j65qg\" (UniqueName: \"kubernetes.io/projected/b448b273-61d9-4418-b752-30f6e6556217-kube-api-access-j65qg\") pod \"machine-config-server-2266x\" (UID: \"b448b273-61d9-4418-b752-30f6e6556217\") " pod="openshift-machine-config-operator/machine-config-server-2266x" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.813723 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c0af2f6a-1320-4a72-9bfc-684d0f666756-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2jcd2\" (UID: \"c0af2f6a-1320-4a72-9bfc-684d0f666756\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2jcd2" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.813963 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/54c65dfa-5787-4f1f-979c-9e5daf517650-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.813988 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/19abbd38-da74-4e94-94c5-9461701ac88d-plugins-dir\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.814040 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80167c35-d69a-44bc-8757-87b7ebf137ac-metrics-tls\") pod \"ingress-operator-5b745b69d9-2vr2v\" (UID: \"80167c35-d69a-44bc-8757-87b7ebf137ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.814062 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/193435f1-4cc8-488e-9010-5b13024376e7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2c4qm\" (UID: \"193435f1-4cc8-488e-9010-5b13024376e7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.814080 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8794d67a-f5f8-4f02-a5d4-8b1f214355bc-srv-cert\") pod \"catalog-operator-68c6474976-l2r9n\" (UID: \"8794d67a-f5f8-4f02-a5d4-8b1f214355bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.814129 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8794d67a-f5f8-4f02-a5d4-8b1f214355bc-profile-collector-cert\") pod \"catalog-operator-68c6474976-l2r9n\" (UID: \"8794d67a-f5f8-4f02-a5d4-8b1f214355bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.814151 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g5ll\" (UniqueName: \"kubernetes.io/projected/8a03bbd9-1c8c-4382-ad9d-c59e2efc9427-kube-api-access-5g5ll\") pod \"dns-default-99fnd\" (UID: \"8a03bbd9-1c8c-4382-ad9d-c59e2efc9427\") " pod="openshift-dns/dns-default-99fnd" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.814173 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8a984dbf-3bbf-475a-9c5e-a7b98f594eaf-tmpfs\") pod \"packageserver-d55dfcdfc-8p4jr\" (UID: \"8a984dbf-3bbf-475a-9c5e-a7b98f594eaf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.814233 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/347982ab-7ed7-4b00-ae43-03b9b91b5133-srv-cert\") pod \"olm-operator-6b444d44fb-f55t6\" (UID: \"347982ab-7ed7-4b00-ae43-03b9b91b5133\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.814292 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26wk4\" (UniqueName: \"kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-kube-api-access-26wk4\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.814360 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz74t\" (UniqueName: \"kubernetes.io/projected/19abbd38-da74-4e94-94c5-9461701ac88d-kube-api-access-lz74t\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.814862 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80167c35-d69a-44bc-8757-87b7ebf137ac-trusted-ca\") pod \"ingress-operator-5b745b69d9-2vr2v\" (UID: \"80167c35-d69a-44bc-8757-87b7ebf137ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.814987 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a90528ef-02f4-4012-bff1-1da159712d79-config\") pod \"service-ca-operator-777779d784-bgbt2\" (UID: \"a90528ef-02f4-4012-bff1-1da159712d79\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.815036 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a90528ef-02f4-4012-bff1-1da159712d79-serving-cert\") pod \"service-ca-operator-777779d784-bgbt2\" (UID: \"a90528ef-02f4-4012-bff1-1da159712d79\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.815117 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/91f97449-8808-4e97-980b-3452f9cb1983-signing-cabundle\") pod \"service-ca-9c57cc56f-r9gw4\" (UID: \"91f97449-8808-4e97-980b-3452f9cb1983\") " pod="openshift-service-ca/service-ca-9c57cc56f-r9gw4" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.815152 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8a984dbf-3bbf-475a-9c5e-a7b98f594eaf-apiservice-cert\") pod \"packageserver-d55dfcdfc-8p4jr\" (UID: \"8a984dbf-3bbf-475a-9c5e-a7b98f594eaf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.815172 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c234ec14-adc4-4089-b41f-0df1e4010599-cert\") pod \"ingress-canary-9r679\" (UID: \"c234ec14-adc4-4089-b41f-0df1e4010599\") " pod="openshift-ingress-canary/ingress-canary-9r679" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.815190 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlgqs\" (UniqueName: \"kubernetes.io/projected/193435f1-4cc8-488e-9010-5b13024376e7-kube-api-access-vlgqs\") pod \"package-server-manager-789f6589d5-2c4qm\" (UID: \"193435f1-4cc8-488e-9010-5b13024376e7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.818331 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/54c65dfa-5787-4f1f-979c-9e5daf517650-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.819454 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/54c65dfa-5787-4f1f-979c-9e5daf517650-registry-certificates\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.842277 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-s6nss"] Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.844286 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-registry-tls\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.845216 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.856695 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.869524 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/54c65dfa-5787-4f1f-979c-9e5daf517650-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.877499 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-bound-sa-token\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.879387 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26wk4\" (UniqueName: \"kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-kube-api-access-26wk4\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.916320 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:55 crc kubenswrapper[4739]: E1201 15:36:55.918307 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:56.418244131 +0000 UTC m=+118.243990225 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.918453 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpwln\" (UniqueName: \"kubernetes.io/projected/80167c35-d69a-44bc-8757-87b7ebf137ac-kube-api-access-dpwln\") pod \"ingress-operator-5b745b69d9-2vr2v\" (UID: \"80167c35-d69a-44bc-8757-87b7ebf137ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.918486 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/19abbd38-da74-4e94-94c5-9461701ac88d-csi-data-dir\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.918509 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a03bbd9-1c8c-4382-ad9d-c59e2efc9427-config-volume\") pod \"dns-default-99fnd\" (UID: \"8a03bbd9-1c8c-4382-ad9d-c59e2efc9427\") " pod="openshift-dns/dns-default-99fnd" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.918545 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8a984dbf-3bbf-475a-9c5e-a7b98f594eaf-webhook-cert\") pod \"packageserver-d55dfcdfc-8p4jr\" (UID: \"8a984dbf-3bbf-475a-9c5e-a7b98f594eaf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.918584 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8a03bbd9-1c8c-4382-ad9d-c59e2efc9427-metrics-tls\") pod \"dns-default-99fnd\" (UID: \"8a03bbd9-1c8c-4382-ad9d-c59e2efc9427\") " pod="openshift-dns/dns-default-99fnd" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.918617 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f7e22852-525d-453f-a329-c15a1c5977ad-secret-volume\") pod \"collect-profiles-29410050-688rv\" (UID: \"f7e22852-525d-453f-a329-c15a1c5977ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.918647 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwr2q\" (UniqueName: \"kubernetes.io/projected/c234ec14-adc4-4089-b41f-0df1e4010599-kube-api-access-fwr2q\") pod \"ingress-canary-9r679\" (UID: \"c234ec14-adc4-4089-b41f-0df1e4010599\") " pod="openshift-ingress-canary/ingress-canary-9r679" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.918678 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b448b273-61d9-4418-b752-30f6e6556217-node-bootstrap-token\") pod \"machine-config-server-2266x\" (UID: \"b448b273-61d9-4418-b752-30f6e6556217\") " pod="openshift-machine-config-operator/machine-config-server-2266x" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.918837 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/19abbd38-da74-4e94-94c5-9461701ac88d-csi-data-dir\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.919323 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a03bbd9-1c8c-4382-ad9d-c59e2efc9427-config-volume\") pod \"dns-default-99fnd\" (UID: \"8a03bbd9-1c8c-4382-ad9d-c59e2efc9427\") " pod="openshift-dns/dns-default-99fnd" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.920346 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xrqs\" (UniqueName: \"kubernetes.io/projected/c0af2f6a-1320-4a72-9bfc-684d0f666756-kube-api-access-6xrqs\") pod \"control-plane-machine-set-operator-78cbb6b69f-2jcd2\" (UID: \"c0af2f6a-1320-4a72-9bfc-684d0f666756\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2jcd2" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.920407 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/80167c35-d69a-44bc-8757-87b7ebf137ac-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2vr2v\" (UID: \"80167c35-d69a-44bc-8757-87b7ebf137ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.921718 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm"] Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.928130 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f7e22852-525d-453f-a329-c15a1c5977ad-secret-volume\") pod \"collect-profiles-29410050-688rv\" (UID: \"f7e22852-525d-453f-a329-c15a1c5977ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.928153 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8a03bbd9-1c8c-4382-ad9d-c59e2efc9427-metrics-tls\") pod \"dns-default-99fnd\" (UID: \"8a03bbd9-1c8c-4382-ad9d-c59e2efc9427\") " pod="openshift-dns/dns-default-99fnd" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.928584 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8a984dbf-3bbf-475a-9c5e-a7b98f594eaf-webhook-cert\") pod \"packageserver-d55dfcdfc-8p4jr\" (UID: \"8a984dbf-3bbf-475a-9c5e-a7b98f594eaf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.930051 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b448b273-61d9-4418-b752-30f6e6556217-node-bootstrap-token\") pod \"machine-config-server-2266x\" (UID: \"b448b273-61d9-4418-b752-30f6e6556217\") " pod="openshift-machine-config-operator/machine-config-server-2266x" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.930606 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-s2hr5"] Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.920492 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j65qg\" (UniqueName: \"kubernetes.io/projected/b448b273-61d9-4418-b752-30f6e6556217-kube-api-access-j65qg\") pod \"machine-config-server-2266x\" (UID: \"b448b273-61d9-4418-b752-30f6e6556217\") " pod="openshift-machine-config-operator/machine-config-server-2266x" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.930709 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c0af2f6a-1320-4a72-9bfc-684d0f666756-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2jcd2\" (UID: \"c0af2f6a-1320-4a72-9bfc-684d0f666756\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2jcd2" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.930768 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/19abbd38-da74-4e94-94c5-9461701ac88d-plugins-dir\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.930794 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80167c35-d69a-44bc-8757-87b7ebf137ac-metrics-tls\") pod \"ingress-operator-5b745b69d9-2vr2v\" (UID: \"80167c35-d69a-44bc-8757-87b7ebf137ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.930816 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/193435f1-4cc8-488e-9010-5b13024376e7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2c4qm\" (UID: \"193435f1-4cc8-488e-9010-5b13024376e7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.930839 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8794d67a-f5f8-4f02-a5d4-8b1f214355bc-srv-cert\") pod \"catalog-operator-68c6474976-l2r9n\" (UID: \"8794d67a-f5f8-4f02-a5d4-8b1f214355bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.930864 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8794d67a-f5f8-4f02-a5d4-8b1f214355bc-profile-collector-cert\") pod \"catalog-operator-68c6474976-l2r9n\" (UID: \"8794d67a-f5f8-4f02-a5d4-8b1f214355bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.930889 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g5ll\" (UniqueName: \"kubernetes.io/projected/8a03bbd9-1c8c-4382-ad9d-c59e2efc9427-kube-api-access-5g5ll\") pod \"dns-default-99fnd\" (UID: \"8a03bbd9-1c8c-4382-ad9d-c59e2efc9427\") " pod="openshift-dns/dns-default-99fnd" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.930908 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8a984dbf-3bbf-475a-9c5e-a7b98f594eaf-tmpfs\") pod \"packageserver-d55dfcdfc-8p4jr\" (UID: \"8a984dbf-3bbf-475a-9c5e-a7b98f594eaf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.930935 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/347982ab-7ed7-4b00-ae43-03b9b91b5133-srv-cert\") pod \"olm-operator-6b444d44fb-f55t6\" (UID: \"347982ab-7ed7-4b00-ae43-03b9b91b5133\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.930983 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz74t\" (UniqueName: \"kubernetes.io/projected/19abbd38-da74-4e94-94c5-9461701ac88d-kube-api-access-lz74t\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931010 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80167c35-d69a-44bc-8757-87b7ebf137ac-trusted-ca\") pod \"ingress-operator-5b745b69d9-2vr2v\" (UID: \"80167c35-d69a-44bc-8757-87b7ebf137ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931040 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a90528ef-02f4-4012-bff1-1da159712d79-config\") pod \"service-ca-operator-777779d784-bgbt2\" (UID: \"a90528ef-02f4-4012-bff1-1da159712d79\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931059 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a90528ef-02f4-4012-bff1-1da159712d79-serving-cert\") pod \"service-ca-operator-777779d784-bgbt2\" (UID: \"a90528ef-02f4-4012-bff1-1da159712d79\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931088 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/91f97449-8808-4e97-980b-3452f9cb1983-signing-cabundle\") pod \"service-ca-9c57cc56f-r9gw4\" (UID: \"91f97449-8808-4e97-980b-3452f9cb1983\") " pod="openshift-service-ca/service-ca-9c57cc56f-r9gw4" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931103 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c234ec14-adc4-4089-b41f-0df1e4010599-cert\") pod \"ingress-canary-9r679\" (UID: \"c234ec14-adc4-4089-b41f-0df1e4010599\") " pod="openshift-ingress-canary/ingress-canary-9r679" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931120 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8a984dbf-3bbf-475a-9c5e-a7b98f594eaf-apiservice-cert\") pod \"packageserver-d55dfcdfc-8p4jr\" (UID: \"8a984dbf-3bbf-475a-9c5e-a7b98f594eaf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931141 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlgqs\" (UniqueName: \"kubernetes.io/projected/193435f1-4cc8-488e-9010-5b13024376e7-kube-api-access-vlgqs\") pod \"package-server-manager-789f6589d5-2c4qm\" (UID: \"193435f1-4cc8-488e-9010-5b13024376e7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931187 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sjf6\" (UniqueName: \"kubernetes.io/projected/347982ab-7ed7-4b00-ae43-03b9b91b5133-kube-api-access-7sjf6\") pod \"olm-operator-6b444d44fb-f55t6\" (UID: \"347982ab-7ed7-4b00-ae43-03b9b91b5133\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931214 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djr8g\" (UniqueName: \"kubernetes.io/projected/8a984dbf-3bbf-475a-9c5e-a7b98f594eaf-kube-api-access-djr8g\") pod \"packageserver-d55dfcdfc-8p4jr\" (UID: \"8a984dbf-3bbf-475a-9c5e-a7b98f594eaf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931242 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f7e22852-525d-453f-a329-c15a1c5977ad-config-volume\") pod \"collect-profiles-29410050-688rv\" (UID: \"f7e22852-525d-453f-a329-c15a1c5977ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931270 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/19abbd38-da74-4e94-94c5-9461701ac88d-socket-dir\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931294 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/19abbd38-da74-4e94-94c5-9461701ac88d-registration-dir\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931326 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4s66\" (UniqueName: \"kubernetes.io/projected/c5277ea8-3db8-406c-9715-d6ce69e75470-kube-api-access-w4s66\") pod \"migrator-59844c95c7-fbk7w\" (UID: \"c5277ea8-3db8-406c-9715-d6ce69e75470\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fbk7w" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931348 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/347982ab-7ed7-4b00-ae43-03b9b91b5133-profile-collector-cert\") pod \"olm-operator-6b444d44fb-f55t6\" (UID: \"347982ab-7ed7-4b00-ae43-03b9b91b5133\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931364 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b448b273-61d9-4418-b752-30f6e6556217-certs\") pod \"machine-config-server-2266x\" (UID: \"b448b273-61d9-4418-b752-30f6e6556217\") " pod="openshift-machine-config-operator/machine-config-server-2266x" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931399 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ec35359a-8d42-4230-8f34-1dcbee21c8f3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-v2m45\" (UID: \"ec35359a-8d42-4230-8f34-1dcbee21c8f3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-v2m45" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931446 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh9ch\" (UniqueName: \"kubernetes.io/projected/ec35359a-8d42-4230-8f34-1dcbee21c8f3-kube-api-access-mh9ch\") pod \"multus-admission-controller-857f4d67dd-v2m45\" (UID: \"ec35359a-8d42-4230-8f34-1dcbee21c8f3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-v2m45" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.931463 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h6j5\" (UniqueName: \"kubernetes.io/projected/a90528ef-02f4-4012-bff1-1da159712d79-kube-api-access-4h6j5\") pod \"service-ca-operator-777779d784-bgbt2\" (UID: \"a90528ef-02f4-4012-bff1-1da159712d79\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.932340 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/19abbd38-da74-4e94-94c5-9461701ac88d-socket-dir\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.932501 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a90528ef-02f4-4012-bff1-1da159712d79-config\") pod \"service-ca-operator-777779d784-bgbt2\" (UID: \"a90528ef-02f4-4012-bff1-1da159712d79\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.932570 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8a984dbf-3bbf-475a-9c5e-a7b98f594eaf-tmpfs\") pod \"packageserver-d55dfcdfc-8p4jr\" (UID: \"8a984dbf-3bbf-475a-9c5e-a7b98f594eaf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.933356 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/91f97449-8808-4e97-980b-3452f9cb1983-signing-cabundle\") pod \"service-ca-9c57cc56f-r9gw4\" (UID: \"91f97449-8808-4e97-980b-3452f9cb1983\") " pod="openshift-service-ca/service-ca-9c57cc56f-r9gw4" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.933788 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80167c35-d69a-44bc-8757-87b7ebf137ac-trusted-ca\") pod \"ingress-operator-5b745b69d9-2vr2v\" (UID: \"80167c35-d69a-44bc-8757-87b7ebf137ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.934227 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/19abbd38-da74-4e94-94c5-9461701ac88d-registration-dir\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.935481 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tf5k\" (UniqueName: \"kubernetes.io/projected/91f97449-8808-4e97-980b-3452f9cb1983-kube-api-access-2tf5k\") pod \"service-ca-9c57cc56f-r9gw4\" (UID: \"91f97449-8808-4e97-980b-3452f9cb1983\") " pod="openshift-service-ca/service-ca-9c57cc56f-r9gw4" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.935514 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/19abbd38-da74-4e94-94c5-9461701ac88d-mountpoint-dir\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.935567 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.935588 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-775d9\" (UniqueName: \"kubernetes.io/projected/8794d67a-f5f8-4f02-a5d4-8b1f214355bc-kube-api-access-775d9\") pod \"catalog-operator-68c6474976-l2r9n\" (UID: \"8794d67a-f5f8-4f02-a5d4-8b1f214355bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.935620 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/91f97449-8808-4e97-980b-3452f9cb1983-signing-key\") pod \"service-ca-9c57cc56f-r9gw4\" (UID: \"91f97449-8808-4e97-980b-3452f9cb1983\") " pod="openshift-service-ca/service-ca-9c57cc56f-r9gw4" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.935740 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwdcg\" (UniqueName: \"kubernetes.io/projected/f7e22852-525d-453f-a329-c15a1c5977ad-kube-api-access-kwdcg\") pod \"collect-profiles-29410050-688rv\" (UID: \"f7e22852-525d-453f-a329-c15a1c5977ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.936401 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/19abbd38-da74-4e94-94c5-9461701ac88d-mountpoint-dir\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: E1201 15:36:55.936664 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:56.436652067 +0000 UTC m=+118.262398161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.937221 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c234ec14-adc4-4089-b41f-0df1e4010599-cert\") pod \"ingress-canary-9r679\" (UID: \"c234ec14-adc4-4089-b41f-0df1e4010599\") " pod="openshift-ingress-canary/ingress-canary-9r679" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.937529 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/347982ab-7ed7-4b00-ae43-03b9b91b5133-profile-collector-cert\") pod \"olm-operator-6b444d44fb-f55t6\" (UID: \"347982ab-7ed7-4b00-ae43-03b9b91b5133\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.938216 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f7e22852-525d-453f-a329-c15a1c5977ad-config-volume\") pod \"collect-profiles-29410050-688rv\" (UID: \"f7e22852-525d-453f-a329-c15a1c5977ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.938240 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/19abbd38-da74-4e94-94c5-9461701ac88d-plugins-dir\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.938290 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8a984dbf-3bbf-475a-9c5e-a7b98f594eaf-apiservice-cert\") pod \"packageserver-d55dfcdfc-8p4jr\" (UID: \"8a984dbf-3bbf-475a-9c5e-a7b98f594eaf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.938905 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8794d67a-f5f8-4f02-a5d4-8b1f214355bc-srv-cert\") pod \"catalog-operator-68c6474976-l2r9n\" (UID: \"8794d67a-f5f8-4f02-a5d4-8b1f214355bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.941924 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ec35359a-8d42-4230-8f34-1dcbee21c8f3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-v2m45\" (UID: \"ec35359a-8d42-4230-8f34-1dcbee21c8f3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-v2m45" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.941979 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c0af2f6a-1320-4a72-9bfc-684d0f666756-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2jcd2\" (UID: \"c0af2f6a-1320-4a72-9bfc-684d0f666756\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2jcd2" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.942251 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b448b273-61d9-4418-b752-30f6e6556217-certs\") pod \"machine-config-server-2266x\" (UID: \"b448b273-61d9-4418-b752-30f6e6556217\") " pod="openshift-machine-config-operator/machine-config-server-2266x" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.943729 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/347982ab-7ed7-4b00-ae43-03b9b91b5133-srv-cert\") pod \"olm-operator-6b444d44fb-f55t6\" (UID: \"347982ab-7ed7-4b00-ae43-03b9b91b5133\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.946423 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8794d67a-f5f8-4f02-a5d4-8b1f214355bc-profile-collector-cert\") pod \"catalog-operator-68c6474976-l2r9n\" (UID: \"8794d67a-f5f8-4f02-a5d4-8b1f214355bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.946651 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80167c35-d69a-44bc-8757-87b7ebf137ac-metrics-tls\") pod \"ingress-operator-5b745b69d9-2vr2v\" (UID: \"80167c35-d69a-44bc-8757-87b7ebf137ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.947520 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a90528ef-02f4-4012-bff1-1da159712d79-serving-cert\") pod \"service-ca-operator-777779d784-bgbt2\" (UID: \"a90528ef-02f4-4012-bff1-1da159712d79\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.947934 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/193435f1-4cc8-488e-9010-5b13024376e7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2c4qm\" (UID: \"193435f1-4cc8-488e-9010-5b13024376e7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.952002 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/91f97449-8808-4e97-980b-3452f9cb1983-signing-key\") pod \"service-ca-9c57cc56f-r9gw4\" (UID: \"91f97449-8808-4e97-980b-3452f9cb1983\") " pod="openshift-service-ca/service-ca-9c57cc56f-r9gw4" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.983986 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwr2q\" (UniqueName: \"kubernetes.io/projected/c234ec14-adc4-4089-b41f-0df1e4010599-kube-api-access-fwr2q\") pod \"ingress-canary-9r679\" (UID: \"c234ec14-adc4-4089-b41f-0df1e4010599\") " pod="openshift-ingress-canary/ingress-canary-9r679" Dec 01 15:36:55 crc kubenswrapper[4739]: I1201 15:36:55.994095 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-9r679" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.013950 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/80167c35-d69a-44bc-8757-87b7ebf137ac-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2vr2v\" (UID: \"80167c35-d69a-44bc-8757-87b7ebf137ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.023173 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpwln\" (UniqueName: \"kubernetes.io/projected/80167c35-d69a-44bc-8757-87b7ebf137ac-kube-api-access-dpwln\") pod \"ingress-operator-5b745b69d9-2vr2v\" (UID: \"80167c35-d69a-44bc-8757-87b7ebf137ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.032400 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xrqs\" (UniqueName: \"kubernetes.io/projected/c0af2f6a-1320-4a72-9bfc-684d0f666756-kube-api-access-6xrqs\") pod \"control-plane-machine-set-operator-78cbb6b69f-2jcd2\" (UID: \"c0af2f6a-1320-4a72-9bfc-684d0f666756\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2jcd2" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.038367 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:56 crc kubenswrapper[4739]: E1201 15:36:56.038824 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:56.538809597 +0000 UTC m=+118.364555691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.060096 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j65qg\" (UniqueName: \"kubernetes.io/projected/b448b273-61d9-4418-b752-30f6e6556217-kube-api-access-j65qg\") pod \"machine-config-server-2266x\" (UID: \"b448b273-61d9-4418-b752-30f6e6556217\") " pod="openshift-machine-config-operator/machine-config-server-2266x" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.073482 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g5ll\" (UniqueName: \"kubernetes.io/projected/8a03bbd9-1c8c-4382-ad9d-c59e2efc9427-kube-api-access-5g5ll\") pod \"dns-default-99fnd\" (UID: \"8a03bbd9-1c8c-4382-ad9d-c59e2efc9427\") " pod="openshift-dns/dns-default-99fnd" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.099331 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz74t\" (UniqueName: \"kubernetes.io/projected/19abbd38-da74-4e94-94c5-9461701ac88d-kube-api-access-lz74t\") pod \"csi-hostpathplugin-t98c7\" (UID: \"19abbd38-da74-4e94-94c5-9461701ac88d\") " pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.105922 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.116712 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djr8g\" (UniqueName: \"kubernetes.io/projected/8a984dbf-3bbf-475a-9c5e-a7b98f594eaf-kube-api-access-djr8g\") pod \"packageserver-d55dfcdfc-8p4jr\" (UID: \"8a984dbf-3bbf-475a-9c5e-a7b98f594eaf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.139284 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.140198 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.140657 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x" event={"ID":"38613c37-a98c-443b-8aaf-c75fcabc9238","Type":"ContainerStarted","Data":"9e6e3a2a2a41b68a29a8e1ebed2e79f2d300615e44cecf18a3bb0f2ff8a3b3fa"} Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.140698 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-sqfdh"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.140713 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x" event={"ID":"38613c37-a98c-443b-8aaf-c75fcabc9238","Type":"ContainerStarted","Data":"08da0b189881fa82f937f4f51383f631a2643f81a588eabafa819dd4772a76bc"} Dec 01 15:36:56 crc kubenswrapper[4739]: E1201 15:36:56.140773 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:56.640759451 +0000 UTC m=+118.466505545 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.146838 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-6b8xv" event={"ID":"217d2158-0c6e-48d9-9713-5499a54655d9","Type":"ContainerStarted","Data":"9e317c92a7917511823068a24767e60e46b0cc339f54cd8940081605887dbdd6"} Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.146873 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-6b8xv" event={"ID":"217d2158-0c6e-48d9-9713-5499a54655d9","Type":"ContainerStarted","Data":"a7a06748eedf3b65636d3a2f0aed74c81cdfc55d926c9c71d6373f46ecf918d1"} Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.151150 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-s2hr5" event={"ID":"8e9c01d1-d2a8-4d89-afb9-9599728cb909","Type":"ContainerStarted","Data":"f070ed261bc530f9f63d414a6b0a6aea81538efbd523b6aca8d0656866acb89c"} Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.153353 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-s6nss" event={"ID":"67608617-45fd-4c45-87f9-325996855e5e","Type":"ContainerStarted","Data":"ddd9f338ef42e29107af09c2390f5a8fad21872eaf2d78452b8ec90487f49082"} Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.154990 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" event={"ID":"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a","Type":"ContainerStarted","Data":"3b55b7fdc27f3466bd42a44e2b254800d6cb6febb29b79ef944c9604bc8f3cba"} Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.155009 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" event={"ID":"95e6ea21-17fb-4a16-8ddd-e0cbc2a0264a","Type":"ContainerStarted","Data":"b8abec77a4bbdcb578f7e56997e40c9347694725493dad748106612da6cd11f8"} Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.155945 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlgqs\" (UniqueName: \"kubernetes.io/projected/193435f1-4cc8-488e-9010-5b13024376e7-kube-api-access-vlgqs\") pod \"package-server-manager-789f6589d5-2c4qm\" (UID: \"193435f1-4cc8-488e-9010-5b13024376e7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.158785 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sjf6\" (UniqueName: \"kubernetes.io/projected/347982ab-7ed7-4b00-ae43-03b9b91b5133-kube-api-access-7sjf6\") pod \"olm-operator-6b444d44fb-f55t6\" (UID: \"347982ab-7ed7-4b00-ae43-03b9b91b5133\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" Dec 01 15:36:56 crc kubenswrapper[4739]: W1201 15:36:56.175601 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29e4371e_f34c_409b_b9db_64de91ddd2ef.slice/crio-4cf18c92accab80d649915b649e227790c8ec77082ae9da1916cc87ccf1c5e37 WatchSource:0}: Error finding container 4cf18c92accab80d649915b649e227790c8ec77082ae9da1916cc87ccf1c5e37: Status 404 returned error can't find the container with id 4cf18c92accab80d649915b649e227790c8ec77082ae9da1916cc87ccf1c5e37 Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.179691 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2jcd2" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.180347 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w" event={"ID":"42bba742-38b4-4d27-9de8-d5f23fe28e84","Type":"ContainerStarted","Data":"35ec722378dfc6647e2fed6972a8eecd4b7ff22308e01c28715437d037b9feef"} Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.180376 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w" event={"ID":"42bba742-38b4-4d27-9de8-d5f23fe28e84","Type":"ContainerStarted","Data":"5c764b8aace137a09c89f54d04af4c744d57cbe1406a9b131cf99f19c72c665b"} Dec 01 15:36:56 crc kubenswrapper[4739]: W1201 15:36:56.183166 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1878674e_9147_4b42_b95a_6c5a82ff5573.slice/crio-b6a6909899540ca73cd6f6ac7e72b10fbf930d3f3712afb42e00bd9032012d8d WatchSource:0}: Error finding container b6a6909899540ca73cd6f6ac7e72b10fbf930d3f3712afb42e00bd9032012d8d: Status 404 returned error can't find the container with id b6a6909899540ca73cd6f6ac7e72b10fbf930d3f3712afb42e00bd9032012d8d Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.183193 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-twbgk" event={"ID":"80a392a8-e94f-4dfa-aefa-3dd05e1051c5","Type":"ContainerStarted","Data":"504b125f1d87f9c00ca59c916ee62fe64fb52bf5b0d61e5aad8e7e3fa00fa4c6"} Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.185232 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" event={"ID":"351ebc2b-c707-4aec-8f6b-e79af328fdf5","Type":"ContainerStarted","Data":"c69859cf51e537b775e7b043491536bffdc9a65317034ac486f5b3aabd7a3e26"} Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.189950 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4s66\" (UniqueName: \"kubernetes.io/projected/c5277ea8-3db8-406c-9715-d6ce69e75470-kube-api-access-w4s66\") pod \"migrator-59844c95c7-fbk7w\" (UID: \"c5277ea8-3db8-406c-9715-d6ce69e75470\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fbk7w" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.192269 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.197536 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm" event={"ID":"2720c151-cb93-4622-ad01-379f5f4b0466","Type":"ContainerStarted","Data":"fb0d39836de2b44e728d819717052323c0251f2d276e43553a2ef796b19c0aa8"} Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.209365 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.209683 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh9ch\" (UniqueName: \"kubernetes.io/projected/ec35359a-8d42-4230-8f34-1dcbee21c8f3-kube-api-access-mh9ch\") pod \"multus-admission-controller-857f4d67dd-v2m45\" (UID: \"ec35359a-8d42-4230-8f34-1dcbee21c8f3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-v2m45" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.228721 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.228799 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h6j5\" (UniqueName: \"kubernetes.io/projected/a90528ef-02f4-4012-bff1-1da159712d79-kube-api-access-4h6j5\") pod \"service-ca-operator-777779d784-bgbt2\" (UID: \"a90528ef-02f4-4012-bff1-1da159712d79\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.241859 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:56 crc kubenswrapper[4739]: E1201 15:36:56.243524 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:56.743503678 +0000 UTC m=+118.569249772 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.244965 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.249338 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.249517 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.251308 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tf5k\" (UniqueName: \"kubernetes.io/projected/91f97449-8808-4e97-980b-3452f9cb1983-kube-api-access-2tf5k\") pod \"service-ca-9c57cc56f-r9gw4\" (UID: \"91f97449-8808-4e97-980b-3452f9cb1983\") " pod="openshift-service-ca/service-ca-9c57cc56f-r9gw4" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.254659 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-775d9\" (UniqueName: \"kubernetes.io/projected/8794d67a-f5f8-4f02-a5d4-8b1f214355bc-kube-api-access-775d9\") pod \"catalog-operator-68c6474976-l2r9n\" (UID: \"8794d67a-f5f8-4f02-a5d4-8b1f214355bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.255388 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fbk7w" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.271974 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2266x" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.273768 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l7mfx"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.275163 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.285223 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwdcg\" (UniqueName: \"kubernetes.io/projected/f7e22852-525d-453f-a329-c15a1c5977ad-kube-api-access-kwdcg\") pod \"collect-profiles-29410050-688rv\" (UID: \"f7e22852-525d-453f-a329-c15a1c5977ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.285762 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-t98c7" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.302586 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-99fnd" Dec 01 15:36:56 crc kubenswrapper[4739]: W1201 15:36:56.336079 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9cbb8f14_1426_4c22_8ccb_f81dbdc8f175.slice/crio-85a045d8bd0d497ded9d38ad1ddf0c4032823afd409fcc7732d2fdbe1e82bac3 WatchSource:0}: Error finding container 85a045d8bd0d497ded9d38ad1ddf0c4032823afd409fcc7732d2fdbe1e82bac3: Status 404 returned error can't find the container with id 85a045d8bd0d497ded9d38ad1ddf0c4032823afd409fcc7732d2fdbe1e82bac3 Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.350346 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:56 crc kubenswrapper[4739]: E1201 15:36:56.350945 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:56.850919874 +0000 UTC m=+118.676665978 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.382475 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mjqqv"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.425946 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.426108 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-b9xb8"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.430715 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-w9l4v"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.438538 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.442737 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:36:56 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:36:56 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:36:56 crc kubenswrapper[4739]: healthz check failed Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.442975 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.453606 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:56 crc kubenswrapper[4739]: E1201 15:36:56.454050 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:56.954029313 +0000 UTC m=+118.779775407 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.485699 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-v2m45" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.505543 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.517330 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.517803 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.534261 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-r9gw4" Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.555457 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:56 crc kubenswrapper[4739]: E1201 15:36:56.555782 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:57.05577095 +0000 UTC m=+118.881517044 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.600244 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.600306 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-zf9xg"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.657052 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:56 crc kubenswrapper[4739]: E1201 15:36:56.657566 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:57.157548649 +0000 UTC m=+118.983294733 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.677110 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.680572 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5dxtl"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.682266 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g"] Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.758621 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:56 crc kubenswrapper[4739]: E1201 15:36:56.759796 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:57.259778661 +0000 UTC m=+119.085524755 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.864829 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:56 crc kubenswrapper[4739]: E1201 15:36:56.865201 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:57.365184418 +0000 UTC m=+119.190930512 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.970280 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:56 crc kubenswrapper[4739]: E1201 15:36:56.970772 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:57.470755688 +0000 UTC m=+119.296501782 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:56 crc kubenswrapper[4739]: I1201 15:36:56.985906 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-9r679"] Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.072157 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:57 crc kubenswrapper[4739]: E1201 15:36:57.072916 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:57.572874958 +0000 UTC m=+119.398621052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.073227 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:57 crc kubenswrapper[4739]: E1201 15:36:57.074077 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:57.574057392 +0000 UTC m=+119.399803486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.124438 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v"] Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.128111 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2jcd2"] Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.132519 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr"] Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.174531 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:57 crc kubenswrapper[4739]: E1201 15:36:57.174827 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:57.674811401 +0000 UTC m=+119.500557495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.277216 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:57 crc kubenswrapper[4739]: E1201 15:36:57.277938 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:57.777924989 +0000 UTC m=+119.603671083 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.356686 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" event={"ID":"8d2be141-5a5f-4bde-b56f-83ea18ff8e99","Type":"ContainerStarted","Data":"c5b05c39b411c73ac2581ad0048bbec666c661b247395f0553676d722ec4a8c6"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.370181 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw" event={"ID":"7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b","Type":"ContainerStarted","Data":"aecca70b45fb80b0aaeed2049f3a67a3d1a0f693ae2e7025b22b668c29db18ac"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.378472 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:57 crc kubenswrapper[4739]: E1201 15:36:57.378791 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:57.87877811 +0000 UTC m=+119.704524204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.406909 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" event={"ID":"f97c9c68-21cb-4f92-b725-3130f00005c8","Type":"ContainerStarted","Data":"64772ff89238dbd4f82a1926a37a8e1801efa84953d6f832b379e071f9ed8991"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.408154 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pmd7x" podStartSLOduration=99.408130941 podStartE2EDuration="1m39.408130941s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:57.364054664 +0000 UTC m=+119.189800758" watchObservedRunningTime="2025-12-01 15:36:57.408130941 +0000 UTC m=+119.233877035" Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.415804 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z" event={"ID":"1878674e-9147-4b42-b95a-6c5a82ff5573","Type":"ContainerStarted","Data":"b5815525dc80cdb5e70f222ea299f2dcd2cc8d269e9c216156e384bff3045721"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.415866 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z" event={"ID":"1878674e-9147-4b42-b95a-6c5a82ff5573","Type":"ContainerStarted","Data":"b6a6909899540ca73cd6f6ac7e72b10fbf930d3f3712afb42e00bd9032012d8d"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.441047 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" event={"ID":"351ebc2b-c707-4aec-8f6b-e79af328fdf5","Type":"ContainerStarted","Data":"92bef7a1b7ee7319c04ff4bc0f351266ab4d548489f077e113078ea824e7b647"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.441130 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" event={"ID":"351ebc2b-c707-4aec-8f6b-e79af328fdf5","Type":"ContainerStarted","Data":"92fc61887823799b12fc39359b3ffc77fcca271d590d056c4cdaca0e837d789b"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.443319 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-85h9w" podStartSLOduration=99.443306494 podStartE2EDuration="1m39.443306494s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:57.427378152 +0000 UTC m=+119.253124266" watchObservedRunningTime="2025-12-01 15:36:57.443306494 +0000 UTC m=+119.269052588" Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.452026 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:36:57 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:36:57 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:36:57 crc kubenswrapper[4739]: healthz check failed Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.453762 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5dxtl" event={"ID":"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3","Type":"ContainerStarted","Data":"429370f7b95781a4e5d1080dd48ade1fd9afe8e4b4a8a4f40c7144d2cd052726"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.456100 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.481005 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:57 crc kubenswrapper[4739]: E1201 15:36:57.485650 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:57.98563372 +0000 UTC m=+119.811379814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.517472 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-6b8xv" podStartSLOduration=99.517450784 podStartE2EDuration="1m39.517450784s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:57.474867691 +0000 UTC m=+119.300613795" watchObservedRunningTime="2025-12-01 15:36:57.517450784 +0000 UTC m=+119.343196878" Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.548151 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" event={"ID":"c886dc3f-4a59-4e31-a900-2c000427c7a9","Type":"ContainerStarted","Data":"9e96c9f14e76bed19291d59816a8b5cdacaa237f09ee9a8a127eec2a381df6c8"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.548542 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.573953 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2"] Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.574088 4739 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-5pnp5 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.574120 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" podUID="c886dc3f-4a59-4e31-a900-2c000427c7a9" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.590134 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:57 crc kubenswrapper[4739]: E1201 15:36:57.591513 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:58.09149862 +0000 UTC m=+119.917244714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.613152 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fbk7w"] Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.650467 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" event={"ID":"29e4371e-f34c-409b-b9db-64de91ddd2ef","Type":"ContainerStarted","Data":"33620a090bc71f0ac798a3845876cab6acba616c1b15254e6f07562dab507ff1"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.650504 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" event={"ID":"29e4371e-f34c-409b-b9db-64de91ddd2ef","Type":"ContainerStarted","Data":"4cf18c92accab80d649915b649e227790c8ec77082ae9da1916cc87ccf1c5e37"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.676339 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g" event={"ID":"12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b","Type":"ContainerStarted","Data":"476cc7be98b57b964d5ddbebe27a05959021e5b3e5711c55d5ab8b215bd6f9ec"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.681246 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2266x" event={"ID":"b448b273-61d9-4418-b752-30f6e6556217","Type":"ContainerStarted","Data":"d119ce0dde800314d668d372d9a36459af6a190b135b20ae34c71738397c7c31"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.687007 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm"] Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.691980 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:57 crc kubenswrapper[4739]: E1201 15:36:57.692538 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:58.192519406 +0000 UTC m=+120.018265500 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.706656 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl" event={"ID":"0234bac5-34c7-4ef3-93f2-4673ebf78d7a","Type":"ContainerStarted","Data":"5af8e5c6d898ef79b0654308c3c100bc2cfe73ce7a67ea71c5dee1068064e926"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.726506 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" event={"ID":"68d2f0b1-e6a9-4133-aa4d-19c40db638d2","Type":"ContainerStarted","Data":"578fcfba26ae6ce4bd0eaeae95ac64e1c9a893362c19d9619ff2f7601c2d2b25"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.799237 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:57 crc kubenswrapper[4739]: E1201 15:36:57.799873 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:58.29985174 +0000 UTC m=+120.125597834 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.861378 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv"] Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.865717 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-sqfdh" event={"ID":"b2995c47-f228-411e-857d-c2830c577ffc","Type":"ContainerStarted","Data":"d29286b5d4d53a99975fd1bbb36094020bd8fe1f0d8991dcda1f8bd7f3dd7b85"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.865783 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-sqfdh" event={"ID":"b2995c47-f228-411e-857d-c2830c577ffc","Type":"ContainerStarted","Data":"fc3858c2a6b8f1e31924871ce5e8d1d85fb90fe28f93fda0d9fa03eca10a2ac5"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.901141 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:57 crc kubenswrapper[4739]: E1201 15:36:57.902796 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:58.402781183 +0000 UTC m=+120.228527277 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.918014 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t98c7"] Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.930689 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n"] Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.930921 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" event={"ID":"6d0c9b6a-66ba-4667-a6ea-833b12934c09","Type":"ContainerStarted","Data":"7b9fbc9606900eb9c759cf43048625bdb7557dc75928559481df49556051f993"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.930946 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" event={"ID":"6d0c9b6a-66ba-4667-a6ea-833b12934c09","Type":"ContainerStarted","Data":"d161a6cbe6ec56c0de8cadad0bc8fcc3ab3c56d36b3709b1d83e71689b75642b"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.941360 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-j47hk" podStartSLOduration=99.941337447 podStartE2EDuration="1m39.941337447s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:57.930486944 +0000 UTC m=+119.756233038" watchObservedRunningTime="2025-12-01 15:36:57.941337447 +0000 UTC m=+119.767083541" Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.941666 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-r9gw4"] Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.968225 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" event={"ID":"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175","Type":"ContainerStarted","Data":"c17daf213ad6ae6b3ce703966eaf27983cec45d5eff97818ffb41d64f349f794"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.968265 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" event={"ID":"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175","Type":"ContainerStarted","Data":"85a045d8bd0d497ded9d38ad1ddf0c4032823afd409fcc7732d2fdbe1e82bac3"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.969039 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.991644 4739 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-l7mfx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.991740 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" podUID="9cbb8f14-1426-4c22-8ccb-f81dbdc8f175" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.992995 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" event={"ID":"6cec3a69-7dbc-4781-a800-452a7e740adc","Type":"ContainerStarted","Data":"f3904bcc1368a32b4ba00c515cd23cf4c5614484b366eda221c7a05b8437cc59"} Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.993458 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.995347 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-v2m45"] Dec 01 15:36:57 crc kubenswrapper[4739]: I1201 15:36:57.996859 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-99fnd"] Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.000207 4739 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-b9xb8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.000249 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" podUID="6cec3a69-7dbc-4781-a800-452a7e740adc" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.002010 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:58 crc kubenswrapper[4739]: E1201 15:36:58.002360 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:58.502346306 +0000 UTC m=+120.328092400 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.006539 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-s6nss" event={"ID":"67608617-45fd-4c45-87f9-325996855e5e","Type":"ContainerStarted","Data":"88c0abd9f7a3143a1345d17abfa2b9c894610d570981dda48af70b6ed1c88f93"} Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.007109 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-s6nss" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.015995 4739 patch_prober.go:28] interesting pod/downloads-7954f5f757-s6nss container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.016043 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-s6nss" podUID="67608617-45fd-4c45-87f9-325996855e5e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.039396 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-twbgk" event={"ID":"80a392a8-e94f-4dfa-aefa-3dd05e1051c5","Type":"ContainerStarted","Data":"0a64c6d41bbd23f76c6a9d3d59c15095aeb4244c19e71ba7a22c09f4c4c2a725"} Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.040023 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-twbgk" event={"ID":"80a392a8-e94f-4dfa-aefa-3dd05e1051c5","Type":"ContainerStarted","Data":"bae685c435f8a14149f6ae3c353e89c15baf6476c00fac70d09c9261e0549b17"} Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.087541 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-cs6j4" podStartSLOduration=100.087523642 podStartE2EDuration="1m40.087523642s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:58.066954733 +0000 UTC m=+119.892700827" watchObservedRunningTime="2025-12-01 15:36:58.087523642 +0000 UTC m=+119.913269736" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.087990 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" event={"ID":"68db36cd-270c-4550-8642-44f72796a198","Type":"ContainerStarted","Data":"6aeaa3b310e9fe415be2cfd752daac59dfeae19057858501345d74eeb7398002"} Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.105803 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.106429 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-48dh8" podStartSLOduration=100.106396693 podStartE2EDuration="1m40.106396693s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:58.091367256 +0000 UTC m=+119.917113350" watchObservedRunningTime="2025-12-01 15:36:58.106396693 +0000 UTC m=+119.932142787" Dec 01 15:36:58 crc kubenswrapper[4739]: E1201 15:36:58.110862 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:58.610848424 +0000 UTC m=+120.436594518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.118143 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" event={"ID":"840099a9-3ba2-4c14-bcac-de2dcf5c5d95","Type":"ContainerStarted","Data":"901a5489cc6fb1ae716c95d51a6ed8f16b05fc791666b45b6f1f741b2700b48a"} Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.118839 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.130599 4739 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-mjqqv container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" start-of-body= Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.130649 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" podUID="840099a9-3ba2-4c14-bcac-de2dcf5c5d95" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.140987 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-s2hr5" event={"ID":"8e9c01d1-d2a8-4d89-afb9-9599728cb909","Type":"ContainerStarted","Data":"70d21fb4d7d5466cf3dd09d93b9eb87bfaf90c6ca9fbd17d5309da8f37e86e1b"} Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.141849 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6"] Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.141876 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.142638 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-s6nss" podStartSLOduration=100.142617747 podStartE2EDuration="1m40.142617747s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:58.140879565 +0000 UTC m=+119.966625669" watchObservedRunningTime="2025-12-01 15:36:58.142617747 +0000 UTC m=+119.968363841" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.149890 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm" event={"ID":"2720c151-cb93-4622-ad01-379f5f4b0466","Type":"ContainerStarted","Data":"72e943547228e5c13ce75e10e3bed66258f1fdeef284424b988975372efbab6b"} Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.170768 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6bt2z" podStartSLOduration=100.170750901 podStartE2EDuration="1m40.170750901s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:58.167120293 +0000 UTC m=+119.992866387" watchObservedRunningTime="2025-12-01 15:36:58.170750901 +0000 UTC m=+119.996496995" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.175480 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-s2hr5" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.199717 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" podStartSLOduration=100.199701019 podStartE2EDuration="1m40.199701019s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:58.198458273 +0000 UTC m=+120.024204367" watchObservedRunningTime="2025-12-01 15:36:58.199701019 +0000 UTC m=+120.025447113" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.219779 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:58 crc kubenswrapper[4739]: E1201 15:36:58.220984 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:58.720969511 +0000 UTC m=+120.546715605 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.276021 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" podStartSLOduration=100.276002733 podStartE2EDuration="1m40.276002733s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:58.233042708 +0000 UTC m=+120.058788802" watchObservedRunningTime="2025-12-01 15:36:58.276002733 +0000 UTC m=+120.101748827" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.324020 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:58 crc kubenswrapper[4739]: E1201 15:36:58.334379 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:58.834359773 +0000 UTC m=+120.660105867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.348253 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-twbgk" podStartSLOduration=100.348234775 podStartE2EDuration="1m40.348234775s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:58.348032859 +0000 UTC m=+120.173778943" watchObservedRunningTime="2025-12-01 15:36:58.348234775 +0000 UTC m=+120.173980869" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.349603 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mnwj9" podStartSLOduration=100.349597786 podStartE2EDuration="1m40.349597786s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:58.311381362 +0000 UTC m=+120.137127456" watchObservedRunningTime="2025-12-01 15:36:58.349597786 +0000 UTC m=+120.175343880" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.427711 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:58 crc kubenswrapper[4739]: E1201 15:36:58.428277 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:58.928256299 +0000 UTC m=+120.754002393 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.429903 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" podStartSLOduration=100.429892667 podStartE2EDuration="1m40.429892667s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:58.388942733 +0000 UTC m=+120.214688827" watchObservedRunningTime="2025-12-01 15:36:58.429892667 +0000 UTC m=+120.255638761" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.446106 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:36:58 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:36:58 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:36:58 crc kubenswrapper[4739]: healthz check failed Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.446192 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.467885 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-s2hr5" podStartSLOduration=100.467863563 podStartE2EDuration="1m40.467863563s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:58.432921737 +0000 UTC m=+120.258667831" watchObservedRunningTime="2025-12-01 15:36:58.467863563 +0000 UTC m=+120.293609657" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.515810 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" podStartSLOduration=100.515784414 podStartE2EDuration="1m40.515784414s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:58.47414822 +0000 UTC m=+120.299894324" watchObservedRunningTime="2025-12-01 15:36:58.515784414 +0000 UTC m=+120.341530508" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.530714 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:58 crc kubenswrapper[4739]: E1201 15:36:58.531029 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:59.031011247 +0000 UTC m=+120.856757341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.601281 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-h7kkm" podStartSLOduration=100.60124774 podStartE2EDuration="1m40.60124774s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:58.547571638 +0000 UTC m=+120.373317722" watchObservedRunningTime="2025-12-01 15:36:58.60124774 +0000 UTC m=+120.426993834" Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.634142 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:58 crc kubenswrapper[4739]: E1201 15:36:58.634658 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:59.13464179 +0000 UTC m=+120.960387884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.742105 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:58 crc kubenswrapper[4739]: E1201 15:36:58.743134 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:59.243118378 +0000 UTC m=+121.068864472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.846059 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:58 crc kubenswrapper[4739]: E1201 15:36:58.846497 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:59.346480773 +0000 UTC m=+121.172226867 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:58 crc kubenswrapper[4739]: I1201 15:36:58.957299 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:58 crc kubenswrapper[4739]: E1201 15:36:58.957854 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:59.457839246 +0000 UTC m=+121.283585340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.059223 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:59 crc kubenswrapper[4739]: E1201 15:36:59.059523 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:59.559501472 +0000 UTC m=+121.385247566 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.160836 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:59 crc kubenswrapper[4739]: E1201 15:36:59.161168 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:59.661157057 +0000 UTC m=+121.486903151 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.178358 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" event={"ID":"80167c35-d69a-44bc-8757-87b7ebf137ac","Type":"ContainerStarted","Data":"c09863cde1c455a9e1ce0bf0f2d68756dcedb692c8fafe74f69cb64a33495d55"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.178893 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" event={"ID":"80167c35-d69a-44bc-8757-87b7ebf137ac","Type":"ContainerStarted","Data":"bed0907566d3e8d5db57d21d6f9ff548cf22b3d1ab0709badf3d4f5ba6262231"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.178905 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" event={"ID":"80167c35-d69a-44bc-8757-87b7ebf137ac","Type":"ContainerStarted","Data":"4270c0da2f32ed3299bd7644ba7136e837434060ba50f59c71e8521f8e1ff179"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.192012 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" event={"ID":"347982ab-7ed7-4b00-ae43-03b9b91b5133","Type":"ContainerStarted","Data":"9916292f37bcd213c8805b4ec7866ec8fc0fea59137dc8301d74822e076f45d8"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.200641 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" event={"ID":"347982ab-7ed7-4b00-ae43-03b9b91b5133","Type":"ContainerStarted","Data":"ed152be2e1729e10e38bac7754e1e81d069bfc2efeef95b1b8b0e793927efade"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.200665 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.203754 4739 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-f55t6 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.203807 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" podUID="347982ab-7ed7-4b00-ae43-03b9b91b5133" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.224079 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" event={"ID":"f7e22852-525d-453f-a329-c15a1c5977ad","Type":"ContainerStarted","Data":"f2a901b2ff5fe1d8f99939b2816c630f4153860853cf1793ef0f32d733a7cea1"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.224138 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" event={"ID":"f7e22852-525d-453f-a329-c15a1c5977ad","Type":"ContainerStarted","Data":"6cd5ff7693a3e2323b95972f82d4ec46c9b5381541c988c4dbaf6021d4d2749e"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.262288 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:59 crc kubenswrapper[4739]: E1201 15:36:59.262959 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:59.762937695 +0000 UTC m=+121.588683779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.277149 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" event={"ID":"f97c9c68-21cb-4f92-b725-3130f00005c8","Type":"ContainerStarted","Data":"6cbc00ac86a97c4eca2ce5f29066f4e30c02515ab2e12fed57c4be6764fcacd0"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.343997 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" event={"ID":"6cec3a69-7dbc-4781-a800-452a7e740adc","Type":"ContainerStarted","Data":"e29a251c5743ed9b8763ff2af19e9f7b99969df8f44297d1db12abc7fbecb33b"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.364854 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:59 crc kubenswrapper[4739]: E1201 15:36:59.366555 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:36:59.866543618 +0000 UTC m=+121.692289712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.369106 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" podStartSLOduration=101.369067334 podStartE2EDuration="1m41.369067334s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:59.344601438 +0000 UTC m=+121.170347532" watchObservedRunningTime="2025-12-01 15:36:59.369067334 +0000 UTC m=+121.194813428" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.369323 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vr2v" podStartSLOduration=101.369317491 podStartE2EDuration="1m41.369317491s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:59.279678652 +0000 UTC m=+121.105424746" watchObservedRunningTime="2025-12-01 15:36:59.369317491 +0000 UTC m=+121.195063585" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.382294 4739 generic.go:334] "Generic (PLEG): container finished" podID="68db36cd-270c-4550-8642-44f72796a198" containerID="9dbb622fdc8077d47ee94f04b355c32d8e01d410720477ccf4719d2feb3c7c0e" exitCode=0 Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.382397 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" event={"ID":"68db36cd-270c-4550-8642-44f72796a198","Type":"ContainerStarted","Data":"96499243fb388f15a83e6ecafcbe4cb3faaa9e4ea2e4f61c5ac7f507a797e1b6"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.382444 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" event={"ID":"68db36cd-270c-4550-8642-44f72796a198","Type":"ContainerDied","Data":"9dbb622fdc8077d47ee94f04b355c32d8e01d410720477ccf4719d2feb3c7c0e"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.382550 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.460775 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:36:59 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:36:59 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:36:59 crc kubenswrapper[4739]: healthz check failed Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.460824 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.468671 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:59 crc kubenswrapper[4739]: E1201 15:36:59.469819 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:36:59.969798292 +0000 UTC m=+121.795544386 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.470749 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm" event={"ID":"193435f1-4cc8-488e-9010-5b13024376e7","Type":"ContainerStarted","Data":"f0a61f6cb752cbac2bdeea6114af991118c9490bb2d8ade58f27d369c9060d13"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.470788 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm" event={"ID":"193435f1-4cc8-488e-9010-5b13024376e7","Type":"ContainerStarted","Data":"b7e2ba77759592808a67141b6e878562fee9a274f2d5b2d771169e9deec81e8a"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.471313 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.484584 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-zf9xg" podStartSLOduration=101.484568669 podStartE2EDuration="1m41.484568669s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:59.415434379 +0000 UTC m=+121.241180473" watchObservedRunningTime="2025-12-01 15:36:59.484568669 +0000 UTC m=+121.310314763" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.487897 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" event={"ID":"840099a9-3ba2-4c14-bcac-de2dcf5c5d95","Type":"ContainerStarted","Data":"d9b7ff277dd4af7f73739de635e00cfddfcbb79d18a745ae037985af277022f8"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.535712 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.576496 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" podStartSLOduration=101.576478655 podStartE2EDuration="1m41.576478655s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:59.485719744 +0000 UTC m=+121.311465838" watchObservedRunningTime="2025-12-01 15:36:59.576478655 +0000 UTC m=+121.402224749" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.577802 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm" podStartSLOduration=101.577796585 podStartE2EDuration="1m41.577796585s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:59.548888347 +0000 UTC m=+121.374634441" watchObservedRunningTime="2025-12-01 15:36:59.577796585 +0000 UTC m=+121.403542679" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.578577 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:59 crc kubenswrapper[4739]: E1201 15:36:59.578914 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:00.078902588 +0000 UTC m=+121.904648682 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.637010 4739 generic.go:334] "Generic (PLEG): container finished" podID="68d2f0b1-e6a9-4133-aa4d-19c40db638d2" containerID="04ac0fc8ec319289556f70c117946681d130325e209244fb53d83d9b012c4d5e" exitCode=0 Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.637955 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" event={"ID":"68d2f0b1-e6a9-4133-aa4d-19c40db638d2","Type":"ContainerDied","Data":"04ac0fc8ec319289556f70c117946681d130325e209244fb53d83d9b012c4d5e"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.638042 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" event={"ID":"68d2f0b1-e6a9-4133-aa4d-19c40db638d2","Type":"ContainerStarted","Data":"f54346c8bd950227c5e1b4379e9fe61696a5bb8519324ebb031ddbe8241b12ae"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.638125 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.661122 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-r9gw4" event={"ID":"91f97449-8808-4e97-980b-3452f9cb1983","Type":"ContainerStarted","Data":"2d913906158255e70e4af6f25ee2b444d23aa878498c7664bd070f47341d95da"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.661178 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-r9gw4" event={"ID":"91f97449-8808-4e97-980b-3452f9cb1983","Type":"ContainerStarted","Data":"4e811a2a6b927fcd74d8b0d3f4f366b37f357ec8cf853faf4ef635a32cc42eeb"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.671778 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw" event={"ID":"7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b","Type":"ContainerStarted","Data":"0630d539b83118e752bb14a67596b2b9552890be8da298bac695b96792af333e"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.671857 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw" event={"ID":"7a3f8f52-cc0d-47d0-aa6d-2ac325e9168b","Type":"ContainerStarted","Data":"a2826b01c60d04e792323c7eafc340bd9c77832cb36e3a539bcce50256e04cfc"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.682660 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:59 crc kubenswrapper[4739]: E1201 15:36:59.683371 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:00.183345575 +0000 UTC m=+122.009091669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.687483 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2266x" event={"ID":"b448b273-61d9-4418-b752-30f6e6556217","Type":"ContainerStarted","Data":"16e165ebe5e4bec95026dfba9c666ba2c417d8a3f3b72952a889d98a230471d3"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.716609 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2jcd2" event={"ID":"c0af2f6a-1320-4a72-9bfc-684d0f666756","Type":"ContainerStarted","Data":"4ba9c0895cd4ed8df1c65cd16158cec10f875872c07d63191c8115a70f67232c"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.716653 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2jcd2" event={"ID":"c0af2f6a-1320-4a72-9bfc-684d0f666756","Type":"ContainerStarted","Data":"85ab50c0ce8bbdab8596a989f5c42bac6b17f9e2f0a41b381ddfe184bd5af430"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.727786 4739 generic.go:334] "Generic (PLEG): container finished" podID="8d2be141-5a5f-4bde-b56f-83ea18ff8e99" containerID="ece0b4f113c440635b559ab1c402b20658d3d47f8be736627b2eef21406f7d86" exitCode=0 Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.728491 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" event={"ID":"8d2be141-5a5f-4bde-b56f-83ea18ff8e99","Type":"ContainerDied","Data":"ece0b4f113c440635b559ab1c402b20658d3d47f8be736627b2eef21406f7d86"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.729850 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" podStartSLOduration=101.72984069500001 podStartE2EDuration="1m41.729840695s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:59.729191045 +0000 UTC m=+121.554937139" watchObservedRunningTime="2025-12-01 15:36:59.729840695 +0000 UTC m=+121.555586789" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.760566 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-99fnd" event={"ID":"8a03bbd9-1c8c-4382-ad9d-c59e2efc9427","Type":"ContainerStarted","Data":"b947649b43214b6b7bc48932a6a414e89b1322b6ed70c56491731bef14991b0f"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.778579 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-r9gw4" podStartSLOduration=101.778562429 podStartE2EDuration="1m41.778562429s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:59.776503848 +0000 UTC m=+121.602249942" watchObservedRunningTime="2025-12-01 15:36:59.778562429 +0000 UTC m=+121.604308523" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.785110 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:59 crc kubenswrapper[4739]: E1201 15:36:59.786806 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:00.286785843 +0000 UTC m=+122.112531997 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.820006 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-2266x" podStartSLOduration=6.819990458 podStartE2EDuration="6.819990458s" podCreationTimestamp="2025-12-01 15:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:59.818744271 +0000 UTC m=+121.644490365" watchObservedRunningTime="2025-12-01 15:36:59.819990458 +0000 UTC m=+121.645736552" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.829518 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5dxtl" event={"ID":"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3","Type":"ContainerStarted","Data":"051064c8ca5edbb931f1b116e15d032114793f2e260c16fec8b47c7fe3b3fa3c"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.855803 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-v2m45" event={"ID":"ec35359a-8d42-4230-8f34-1dcbee21c8f3","Type":"ContainerStarted","Data":"aaa229da7355870d8d6a908cd56f14cfe4408b96a277ada1b3d5fd267f26d4a3"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.886298 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:36:59 crc kubenswrapper[4739]: E1201 15:36:59.887866 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:00.38782438 +0000 UTC m=+122.213570474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.899213 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:36:59 crc kubenswrapper[4739]: E1201 15:36:59.901469 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:00.401444164 +0000 UTC m=+122.227190258 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.917056 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2jcd2" podStartSLOduration=101.917039767 podStartE2EDuration="1m41.917039767s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:59.916773299 +0000 UTC m=+121.742519383" watchObservedRunningTime="2025-12-01 15:36:59.917039767 +0000 UTC m=+121.742785861" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.924809 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fbk7w" event={"ID":"c5277ea8-3db8-406c-9715-d6ce69e75470","Type":"ContainerStarted","Data":"949b04d4d1a05b609ffc66a9a7c78d607fbbe020400908855fe7ff56049546d1"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.924857 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fbk7w" event={"ID":"c5277ea8-3db8-406c-9715-d6ce69e75470","Type":"ContainerStarted","Data":"063a51f57d9823db3eef7130ce64dafcbb4a6424aec280f750745838490dbf93"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.979797 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qvnkw" podStartSLOduration=101.979779157 podStartE2EDuration="1m41.979779157s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:36:59.959018402 +0000 UTC m=+121.784764496" watchObservedRunningTime="2025-12-01 15:36:59.979779157 +0000 UTC m=+121.805525251" Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.988376 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2" event={"ID":"a90528ef-02f4-4012-bff1-1da159712d79","Type":"ContainerStarted","Data":"df2e3eacbed1bdb6d97a9952769b03a1370d22312ab238473c46046d6f0f8f96"} Dec 01 15:36:59 crc kubenswrapper[4739]: I1201 15:36:59.988452 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2" event={"ID":"a90528ef-02f4-4012-bff1-1da159712d79","Type":"ContainerStarted","Data":"d4d3b397e05ffe3ddc4abba4067fd12ddb6115f93652c3b05effb5087359343f"} Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.004672 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:37:00 crc kubenswrapper[4739]: E1201 15:37:00.006364 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:00.506346176 +0000 UTC m=+122.332092270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.049068 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl" event={"ID":"0234bac5-34c7-4ef3-93f2-4673ebf78d7a","Type":"ContainerStarted","Data":"622f5040f6c7e1704abb785d27b33df5b427794edaa34632acd3d8f97fa4ca1d"} Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.067635 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" event={"ID":"c886dc3f-4a59-4e31-a900-2c000427c7a9","Type":"ContainerStarted","Data":"aec86c2c39bf8605c5471a9dda6f5f4e0828731916faf83a29386ff62e7499db"} Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.085807 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" event={"ID":"8794d67a-f5f8-4f02-a5d4-8b1f214355bc","Type":"ContainerStarted","Data":"75f9dc6e5d67846b9c99b3d0ade778c04e92975763a938d5127716495c15a85f"} Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.085850 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" event={"ID":"8794d67a-f5f8-4f02-a5d4-8b1f214355bc","Type":"ContainerStarted","Data":"3519ec37e1a62b33a5143db735665f591bc67ac2d17f545cd1dcadd55a01fa3d"} Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.086504 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.090930 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.104792 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fbk7w" podStartSLOduration=102.104776095 podStartE2EDuration="1m42.104776095s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:00.103791126 +0000 UTC m=+121.929537220" watchObservedRunningTime="2025-12-01 15:37:00.104776095 +0000 UTC m=+121.930522189" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.105099 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-5dxtl" podStartSLOduration=102.105095375 podStartE2EDuration="1m42.105095375s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:00.029821262 +0000 UTC m=+121.855567356" watchObservedRunningTime="2025-12-01 15:37:00.105095375 +0000 UTC m=+121.930841469" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.106815 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:00 crc kubenswrapper[4739]: E1201 15:37:00.108156 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:00.608140705 +0000 UTC m=+122.433886799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.112599 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" event={"ID":"6d0c9b6a-66ba-4667-a6ea-833b12934c09","Type":"ContainerStarted","Data":"c56fa5134022d043d026a28bd88a2b31c2f84d8f17eb71acee91127fec288825"} Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.124466 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bb5fl" podStartSLOduration=102.124444029 podStartE2EDuration="1m42.124444029s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:00.122568062 +0000 UTC m=+121.948314156" watchObservedRunningTime="2025-12-01 15:37:00.124444029 +0000 UTC m=+121.950190123" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.137685 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-9r679" event={"ID":"c234ec14-adc4-4089-b41f-0df1e4010599","Type":"ContainerStarted","Data":"74eed640ca7e9ea8975e652cd32f4226c8d9539d7840c401b4fa27c09bebb578"} Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.137759 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-9r679" event={"ID":"c234ec14-adc4-4089-b41f-0df1e4010599","Type":"ContainerStarted","Data":"1c5a6e91412a152090b727a9505f8d7412adfad2e4fb5e894f77fcf15bf13b30"} Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.175488 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" event={"ID":"8a984dbf-3bbf-475a-9c5e-a7b98f594eaf","Type":"ContainerStarted","Data":"752ef6247e2cb91f7008d2dfad8e40977f1662746985d275e84916a2fc6edebc"} Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.175534 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" event={"ID":"8a984dbf-3bbf-475a-9c5e-a7b98f594eaf","Type":"ContainerStarted","Data":"bd9f47aa090fde4e205337e6800050e5334e298104c85b5aaa9aa4f194bd562d"} Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.176447 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.193998 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g" event={"ID":"12c7cfeb-0fd0-4de3-bc4c-27d37ce3484b","Type":"ContainerStarted","Data":"013e27b46e161ec39b17531e6e8e1e1a7a0297524e2029b34ac999b47bdf5e97"} Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.204885 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-bgbt2" podStartSLOduration=102.204869384 podStartE2EDuration="1m42.204869384s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:00.203847503 +0000 UTC m=+122.029593597" watchObservedRunningTime="2025-12-01 15:37:00.204869384 +0000 UTC m=+122.030615468" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.210405 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t98c7" event={"ID":"19abbd38-da74-4e94-94c5-9461701ac88d","Type":"ContainerStarted","Data":"0f336008a7c862df4fa0350d7b57ed775f09363da3bb7358230334a479b8080f"} Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.217727 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:37:00 crc kubenswrapper[4739]: E1201 15:37:00.219092 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:00.719075555 +0000 UTC m=+122.544821649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.229322 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.238653 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-sqfdh" event={"ID":"b2995c47-f228-411e-857d-c2830c577ffc","Type":"ContainerStarted","Data":"d91af68e95a74ab83ecca9e0a982bf5804d332bfb49af2e03112bcc61580aee9"} Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.245341 4739 patch_prober.go:28] interesting pod/downloads-7954f5f757-s6nss container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.245378 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-s6nss" podUID="67608617-45fd-4c45-87f9-325996855e5e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.245577 4739 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-l7mfx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.245643 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" podUID="9cbb8f14-1426-4c22-8ccb-f81dbdc8f175" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.254870 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l2r9n" podStartSLOduration=102.254841836 podStartE2EDuration="1m42.254841836s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:00.240315135 +0000 UTC m=+122.066061229" watchObservedRunningTime="2025-12-01 15:37:00.254841836 +0000 UTC m=+122.080587930" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.300724 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jkjv" podStartSLOduration=102.300707177 podStartE2EDuration="1m42.300707177s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:00.299865401 +0000 UTC m=+122.125611505" watchObservedRunningTime="2025-12-01 15:37:00.300707177 +0000 UTC m=+122.126453271" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.319896 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:00 crc kubenswrapper[4739]: E1201 15:37:00.324378 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:00.824358508 +0000 UTC m=+122.650104602 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.363541 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" podStartSLOduration=102.363519429 podStartE2EDuration="1m42.363519429s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:00.341012052 +0000 UTC m=+122.166758146" watchObservedRunningTime="2025-12-01 15:37:00.363519429 +0000 UTC m=+122.189265523" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.366228 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-sqfdh" podStartSLOduration=102.366218819 podStartE2EDuration="1m42.366218819s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:00.361877921 +0000 UTC m=+122.187624015" watchObservedRunningTime="2025-12-01 15:37:00.366218819 +0000 UTC m=+122.191964913" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.423572 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:37:00 crc kubenswrapper[4739]: E1201 15:37:00.456746 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:00.956689273 +0000 UTC m=+122.782435357 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.460847 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-9r679" podStartSLOduration=7.460826286 podStartE2EDuration="7.460826286s" podCreationTimestamp="2025-12-01 15:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:00.404544777 +0000 UTC m=+122.230290871" watchObservedRunningTime="2025-12-01 15:37:00.460826286 +0000 UTC m=+122.286572380" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.471597 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:00 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:00 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:00 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.471677 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.472154 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ndt7g" podStartSLOduration=102.472114611 podStartE2EDuration="1m42.472114611s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:00.464897206 +0000 UTC m=+122.290643300" watchObservedRunningTime="2025-12-01 15:37:00.472114611 +0000 UTC m=+122.297860705" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.531589 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:00 crc kubenswrapper[4739]: E1201 15:37:00.532010 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:01.031995596 +0000 UTC m=+122.857741690 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.634759 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:37:00 crc kubenswrapper[4739]: E1201 15:37:00.635448 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:01.135432044 +0000 UTC m=+122.961178138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.736385 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:00 crc kubenswrapper[4739]: E1201 15:37:00.736741 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:01.236729389 +0000 UTC m=+123.062475483 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.827986 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.837909 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:37:00 crc kubenswrapper[4739]: E1201 15:37:00.838237 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:01.33822352 +0000 UTC m=+123.163969614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:00 crc kubenswrapper[4739]: I1201 15:37:00.939498 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:00 crc kubenswrapper[4739]: E1201 15:37:00.940835 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:01.440823332 +0000 UTC m=+123.266569426 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.040502 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:37:01 crc kubenswrapper[4739]: E1201 15:37:01.040641 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:01.540622872 +0000 UTC m=+123.366368966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.040741 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:01 crc kubenswrapper[4739]: E1201 15:37:01.041046 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:01.541038365 +0000 UTC m=+123.366784459 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.064674 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8p4jr" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.142142 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:37:01 crc kubenswrapper[4739]: E1201 15:37:01.142553 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:01.642539636 +0000 UTC m=+123.468285730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.244180 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:01 crc kubenswrapper[4739]: E1201 15:37:01.244519 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:01.74450355 +0000 UTC m=+123.570249644 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.263564 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t98c7" event={"ID":"19abbd38-da74-4e94-94c5-9461701ac88d","Type":"ContainerStarted","Data":"c2508d057923437d11be276ba81f6cd092da83095ec4f17fbe87582dfd4af21a"} Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.263604 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t98c7" event={"ID":"19abbd38-da74-4e94-94c5-9461701ac88d","Type":"ContainerStarted","Data":"9411b9071f5bfc0685ff3d835f36efb5a29bfa168e705be9d7755b0a395c6127"} Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.270514 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" event={"ID":"8d2be141-5a5f-4bde-b56f-83ea18ff8e99","Type":"ContainerStarted","Data":"cdb4dbfa74e5b0d193b75d76d3683949b9b08a102eac696bd38e51ccaea0d26b"} Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.273480 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" event={"ID":"68db36cd-270c-4550-8642-44f72796a198","Type":"ContainerStarted","Data":"f6d6ac8dce52df6a54c0e0b90e4452bfc3c3f076dc2dc9a86d2eede1b653f308"} Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.275366 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-99fnd" event={"ID":"8a03bbd9-1c8c-4382-ad9d-c59e2efc9427","Type":"ContainerStarted","Data":"e9f88699cb5c8a9f70f87173cf51ea012a47f6cca2646936049ab7ec5b0ed1a9"} Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.275396 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-99fnd" event={"ID":"8a03bbd9-1c8c-4382-ad9d-c59e2efc9427","Type":"ContainerStarted","Data":"15c49c913acf15f38d0c5bc6240a32a9117e6d1370350e5bebbb7a2d751d1612"} Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.276002 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-99fnd" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.277705 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-v2m45" event={"ID":"ec35359a-8d42-4230-8f34-1dcbee21c8f3","Type":"ContainerStarted","Data":"ef8e62bc902ce40f42a92e4fe5ceffec7ce671be56e34ab23cbc71eb7796bfd6"} Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.277743 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-v2m45" event={"ID":"ec35359a-8d42-4230-8f34-1dcbee21c8f3","Type":"ContainerStarted","Data":"0472c00f471446882730cabd1a3708446bae44c0511963026c8af3389fcda506"} Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.283747 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm" event={"ID":"193435f1-4cc8-488e-9010-5b13024376e7","Type":"ContainerStarted","Data":"c0ad6e4c564795e09e4e0d3c6ca2d17eac94fbd3936a3a8997e8bdc6380ae484"} Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.293377 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fbk7w" event={"ID":"c5277ea8-3db8-406c-9715-d6ce69e75470","Type":"ContainerStarted","Data":"07c8f962a945645607e507de7e2aba46f84219872c0980675eb34504f74e7744"} Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.298313 4739 patch_prober.go:28] interesting pod/downloads-7954f5f757-s6nss container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.298352 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-s6nss" podUID="67608617-45fd-4c45-87f9-325996855e5e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.299558 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.300528 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" podStartSLOduration=103.300519582 podStartE2EDuration="1m43.300519582s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:01.300305125 +0000 UTC m=+123.126051219" watchObservedRunningTime="2025-12-01 15:37:01.300519582 +0000 UTC m=+123.126265676" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.308524 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f55t6" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.328482 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" podStartSLOduration=103.328467021 podStartE2EDuration="1m43.328467021s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:01.32710608 +0000 UTC m=+123.152852184" watchObservedRunningTime="2025-12-01 15:37:01.328467021 +0000 UTC m=+123.154213115" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.344963 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:37:01 crc kubenswrapper[4739]: E1201 15:37:01.345094 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:01.845079963 +0000 UTC m=+123.670826057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.345477 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:01 crc kubenswrapper[4739]: E1201 15:37:01.353363 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:01.853350409 +0000 UTC m=+123.679096503 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.366213 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9zvgq" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.450018 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-99fnd" podStartSLOduration=8.449993875 podStartE2EDuration="8.449993875s" podCreationTimestamp="2025-12-01 15:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:01.448195462 +0000 UTC m=+123.273941556" watchObservedRunningTime="2025-12-01 15:37:01.449993875 +0000 UTC m=+123.275739969" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.460575 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.462834 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:01 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:01 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:01 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.462879 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:01 crc kubenswrapper[4739]: E1201 15:37:01.463846 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:01.963828775 +0000 UTC m=+123.789574859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.488851 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-v2m45" podStartSLOduration=103.488830026 podStartE2EDuration="1m43.488830026s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:01.488586709 +0000 UTC m=+123.314332803" watchObservedRunningTime="2025-12-01 15:37:01.488830026 +0000 UTC m=+123.314576120" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.511246 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8vzjc"] Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.512181 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:37:01 crc kubenswrapper[4739]: W1201 15:37:01.534019 4739 reflector.go:561] object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g": failed to list *v1.Secret: secrets "certified-operators-dockercfg-4rs5g" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Dec 01 15:37:01 crc kubenswrapper[4739]: E1201 15:37:01.534072 4739 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-4rs5g\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"certified-operators-dockercfg-4rs5g\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.543037 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8vzjc"] Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.566813 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75fd8\" (UniqueName: \"kubernetes.io/projected/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-kube-api-access-75fd8\") pod \"certified-operators-8vzjc\" (UID: \"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4\") " pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.566861 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.566932 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-utilities\") pod \"certified-operators-8vzjc\" (UID: \"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4\") " pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.566953 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-catalog-content\") pod \"certified-operators-8vzjc\" (UID: \"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4\") " pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:37:01 crc kubenswrapper[4739]: E1201 15:37:01.567269 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:02.067256953 +0000 UTC m=+123.893003047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.656379 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nwqtt"] Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.657240 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.661885 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.667622 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:37:01 crc kubenswrapper[4739]: E1201 15:37:01.667785 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:02.167760583 +0000 UTC m=+123.993506677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.667812 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-catalog-content\") pod \"certified-operators-8vzjc\" (UID: \"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4\") " pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.667878 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75fd8\" (UniqueName: \"kubernetes.io/projected/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-kube-api-access-75fd8\") pod \"certified-operators-8vzjc\" (UID: \"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4\") " pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.667906 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.667977 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-utilities\") pod \"certified-operators-8vzjc\" (UID: \"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4\") " pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.668240 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-catalog-content\") pod \"certified-operators-8vzjc\" (UID: \"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4\") " pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.668321 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-utilities\") pod \"certified-operators-8vzjc\" (UID: \"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4\") " pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:37:01 crc kubenswrapper[4739]: E1201 15:37:01.668473 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:02.168452534 +0000 UTC m=+123.994198698 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.680508 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwqtt"] Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.725276 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75fd8\" (UniqueName: \"kubernetes.io/projected/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-kube-api-access-75fd8\") pod \"certified-operators-8vzjc\" (UID: \"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4\") " pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.768596 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:37:01 crc kubenswrapper[4739]: E1201 15:37:01.768779 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:02.268752579 +0000 UTC m=+124.094498673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.768861 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlwr9\" (UniqueName: \"kubernetes.io/projected/5b9930dd-3336-438d-84fc-3a62b6a86e75-kube-api-access-zlwr9\") pod \"community-operators-nwqtt\" (UID: \"5b9930dd-3336-438d-84fc-3a62b6a86e75\") " pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.768961 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b9930dd-3336-438d-84fc-3a62b6a86e75-catalog-content\") pod \"community-operators-nwqtt\" (UID: \"5b9930dd-3336-438d-84fc-3a62b6a86e75\") " pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.769013 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b9930dd-3336-438d-84fc-3a62b6a86e75-utilities\") pod \"community-operators-nwqtt\" (UID: \"5b9930dd-3336-438d-84fc-3a62b6a86e75\") " pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.828099 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2bhxd"] Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.829354 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.849268 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2bhxd"] Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.870187 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlwr9\" (UniqueName: \"kubernetes.io/projected/5b9930dd-3336-438d-84fc-3a62b6a86e75-kube-api-access-zlwr9\") pod \"community-operators-nwqtt\" (UID: \"5b9930dd-3336-438d-84fc-3a62b6a86e75\") " pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.870254 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b9930dd-3336-438d-84fc-3a62b6a86e75-catalog-content\") pod \"community-operators-nwqtt\" (UID: \"5b9930dd-3336-438d-84fc-3a62b6a86e75\") " pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.870300 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b9930dd-3336-438d-84fc-3a62b6a86e75-utilities\") pod \"community-operators-nwqtt\" (UID: \"5b9930dd-3336-438d-84fc-3a62b6a86e75\") " pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.870329 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:01 crc kubenswrapper[4739]: E1201 15:37:01.870653 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:02.370637781 +0000 UTC m=+124.196383875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.870843 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b9930dd-3336-438d-84fc-3a62b6a86e75-catalog-content\") pod \"community-operators-nwqtt\" (UID: \"5b9930dd-3336-438d-84fc-3a62b6a86e75\") " pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.871017 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b9930dd-3336-438d-84fc-3a62b6a86e75-utilities\") pod \"community-operators-nwqtt\" (UID: \"5b9930dd-3336-438d-84fc-3a62b6a86e75\") " pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.911113 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlwr9\" (UniqueName: \"kubernetes.io/projected/5b9930dd-3336-438d-84fc-3a62b6a86e75-kube-api-access-zlwr9\") pod \"community-operators-nwqtt\" (UID: \"5b9930dd-3336-438d-84fc-3a62b6a86e75\") " pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.968576 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.971572 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:37:01 crc kubenswrapper[4739]: E1201 15:37:01.971785 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:02.47173452 +0000 UTC m=+124.297480614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.971855 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5dkv\" (UniqueName: \"kubernetes.io/projected/4c29aacd-8294-4a12-b340-253e25ebe298-kube-api-access-k5dkv\") pod \"certified-operators-2bhxd\" (UID: \"4c29aacd-8294-4a12-b340-253e25ebe298\") " pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.971960 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c29aacd-8294-4a12-b340-253e25ebe298-utilities\") pod \"certified-operators-2bhxd\" (UID: \"4c29aacd-8294-4a12-b340-253e25ebe298\") " pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.972038 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:01 crc kubenswrapper[4739]: I1201 15:37:01.972289 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c29aacd-8294-4a12-b340-253e25ebe298-catalog-content\") pod \"certified-operators-2bhxd\" (UID: \"4c29aacd-8294-4a12-b340-253e25ebe298\") " pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:37:01 crc kubenswrapper[4739]: E1201 15:37:01.972500 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:02.472479482 +0000 UTC m=+124.298225576 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.057206 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r2t8q"] Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.058102 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.073286 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:37:02 crc kubenswrapper[4739]: E1201 15:37:02.073457 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:02.573415966 +0000 UTC m=+124.399162060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.073665 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c29aacd-8294-4a12-b340-253e25ebe298-utilities\") pod \"certified-operators-2bhxd\" (UID: \"4c29aacd-8294-4a12-b340-253e25ebe298\") " pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.073716 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.073820 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c29aacd-8294-4a12-b340-253e25ebe298-catalog-content\") pod \"certified-operators-2bhxd\" (UID: \"4c29aacd-8294-4a12-b340-253e25ebe298\") " pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.073873 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5dkv\" (UniqueName: \"kubernetes.io/projected/4c29aacd-8294-4a12-b340-253e25ebe298-kube-api-access-k5dkv\") pod \"certified-operators-2bhxd\" (UID: \"4c29aacd-8294-4a12-b340-253e25ebe298\") " pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:37:02 crc kubenswrapper[4739]: E1201 15:37:02.074096 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 15:37:02.574078575 +0000 UTC m=+124.399824669 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kqk8p" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.074125 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c29aacd-8294-4a12-b340-253e25ebe298-utilities\") pod \"certified-operators-2bhxd\" (UID: \"4c29aacd-8294-4a12-b340-253e25ebe298\") " pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.074306 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c29aacd-8294-4a12-b340-253e25ebe298-catalog-content\") pod \"certified-operators-2bhxd\" (UID: \"4c29aacd-8294-4a12-b340-253e25ebe298\") " pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.074937 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r2t8q"] Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.109776 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5dkv\" (UniqueName: \"kubernetes.io/projected/4c29aacd-8294-4a12-b340-253e25ebe298-kube-api-access-k5dkv\") pod \"certified-operators-2bhxd\" (UID: \"4c29aacd-8294-4a12-b340-253e25ebe298\") " pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.125182 4739 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.175014 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.175196 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpjkw\" (UniqueName: \"kubernetes.io/projected/68008673-38d1-477e-8963-1f4d069ba0c7-kube-api-access-dpjkw\") pod \"community-operators-r2t8q\" (UID: \"68008673-38d1-477e-8963-1f4d069ba0c7\") " pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.175235 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68008673-38d1-477e-8963-1f4d069ba0c7-catalog-content\") pod \"community-operators-r2t8q\" (UID: \"68008673-38d1-477e-8963-1f4d069ba0c7\") " pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.175282 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68008673-38d1-477e-8963-1f4d069ba0c7-utilities\") pod \"community-operators-r2t8q\" (UID: \"68008673-38d1-477e-8963-1f4d069ba0c7\") " pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:37:02 crc kubenswrapper[4739]: E1201 15:37:02.175597 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 15:37:02.675538345 +0000 UTC m=+124.501284439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.226996 4739 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-01T15:37:02.125243643Z","Handler":null,"Name":""} Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.253657 4739 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.253707 4739 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.277939 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpjkw\" (UniqueName: \"kubernetes.io/projected/68008673-38d1-477e-8963-1f4d069ba0c7-kube-api-access-dpjkw\") pod \"community-operators-r2t8q\" (UID: \"68008673-38d1-477e-8963-1f4d069ba0c7\") " pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.278001 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68008673-38d1-477e-8963-1f4d069ba0c7-catalog-content\") pod \"community-operators-r2t8q\" (UID: \"68008673-38d1-477e-8963-1f4d069ba0c7\") " pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.278054 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68008673-38d1-477e-8963-1f4d069ba0c7-utilities\") pod \"community-operators-r2t8q\" (UID: \"68008673-38d1-477e-8963-1f4d069ba0c7\") " pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.278090 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.279157 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68008673-38d1-477e-8963-1f4d069ba0c7-catalog-content\") pod \"community-operators-r2t8q\" (UID: \"68008673-38d1-477e-8963-1f4d069ba0c7\") " pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.281837 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68008673-38d1-477e-8963-1f4d069ba0c7-utilities\") pod \"community-operators-r2t8q\" (UID: \"68008673-38d1-477e-8963-1f4d069ba0c7\") " pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.310448 4739 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.310498 4739 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.376628 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpjkw\" (UniqueName: \"kubernetes.io/projected/68008673-38d1-477e-8963-1f4d069ba0c7-kube-api-access-dpjkw\") pod \"community-operators-r2t8q\" (UID: \"68008673-38d1-477e-8963-1f4d069ba0c7\") " pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.391760 4739 generic.go:334] "Generic (PLEG): container finished" podID="f7e22852-525d-453f-a329-c15a1c5977ad" containerID="f2a901b2ff5fe1d8f99939b2816c630f4153860853cf1793ef0f32d733a7cea1" exitCode=0 Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.392463 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" event={"ID":"f7e22852-525d-453f-a329-c15a1c5977ad","Type":"ContainerDied","Data":"f2a901b2ff5fe1d8f99939b2816c630f4153860853cf1793ef0f32d733a7cea1"} Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.398631 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t98c7" event={"ID":"19abbd38-da74-4e94-94c5-9461701ac88d","Type":"ContainerStarted","Data":"2c48014a2fedcc520f4a86c9d1d61b53c4038d3298c10a08565c815ebb4ba5a4"} Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.416272 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kqk8p\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.437992 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.438550 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.442403 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:02 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:02 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:02 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.442466 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.442977 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwqtt"] Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.492202 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.525250 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.671406 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.704581 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kqk8p"] Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.730296 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.731268 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.734917 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.735145 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.738761 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.800046 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be4aaf21-24c7-4500-8151-07af3fab2bcb-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"be4aaf21-24c7-4500-8151-07af3fab2bcb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.800158 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be4aaf21-24c7-4500-8151-07af3fab2bcb-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"be4aaf21-24c7-4500-8151-07af3fab2bcb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.854511 4739 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-marketplace/certified-operators-8vzjc" secret="" err="failed to sync secret cache: timed out waiting for the condition" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.854587 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.904602 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be4aaf21-24c7-4500-8151-07af3fab2bcb-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"be4aaf21-24c7-4500-8151-07af3fab2bcb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.905017 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be4aaf21-24c7-4500-8151-07af3fab2bcb-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"be4aaf21-24c7-4500-8151-07af3fab2bcb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.905097 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be4aaf21-24c7-4500-8151-07af3fab2bcb-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"be4aaf21-24c7-4500-8151-07af3fab2bcb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 15:37:02 crc kubenswrapper[4739]: I1201 15:37:02.939030 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be4aaf21-24c7-4500-8151-07af3fab2bcb-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"be4aaf21-24c7-4500-8151-07af3fab2bcb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.048756 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.052750 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.070051 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.108708 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8vzjc"] Dec 01 15:37:03 crc kubenswrapper[4739]: W1201 15:37:03.116374 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfc6b860_92d2_4c21_b9ec_e7d39b3ad9b4.slice/crio-51269b1fdb9ed1e7e6f938e96692530966b85c93102c20b54ab99293fb64ca9c WatchSource:0}: Error finding container 51269b1fdb9ed1e7e6f938e96692530966b85c93102c20b54ab99293fb64ca9c: Status 404 returned error can't find the container with id 51269b1fdb9ed1e7e6f938e96692530966b85c93102c20b54ab99293fb64ca9c Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.175453 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r2t8q"] Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.353033 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.405736 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vzjc" event={"ID":"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4","Type":"ContainerDied","Data":"dad668596b58be1e97b2b883cf9ee74b65331e3802cb846518cf5df2fd724f73"} Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.406594 4739 generic.go:334] "Generic (PLEG): container finished" podID="cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" containerID="dad668596b58be1e97b2b883cf9ee74b65331e3802cb846518cf5df2fd724f73" exitCode=0 Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.406744 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vzjc" event={"ID":"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4","Type":"ContainerStarted","Data":"51269b1fdb9ed1e7e6f938e96692530966b85c93102c20b54ab99293fb64ca9c"} Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.408164 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2t8q" event={"ID":"68008673-38d1-477e-8963-1f4d069ba0c7","Type":"ContainerStarted","Data":"cbf676fdf19bb60a7c099dcbd720db64727db65ab0a771bf27023922ad5aa588"} Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.409980 4739 generic.go:334] "Generic (PLEG): container finished" podID="5b9930dd-3336-438d-84fc-3a62b6a86e75" containerID="6797e64177018ca5546315738e5e471866e6af1b2b7dcac7cfa1844e81452155" exitCode=0 Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.410224 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwqtt" event={"ID":"5b9930dd-3336-438d-84fc-3a62b6a86e75","Type":"ContainerDied","Data":"6797e64177018ca5546315738e5e471866e6af1b2b7dcac7cfa1844e81452155"} Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.410277 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwqtt" event={"ID":"5b9930dd-3336-438d-84fc-3a62b6a86e75","Type":"ContainerStarted","Data":"79f150d4030b41125f63f7869de679675c37692569b6d7e019a14090f7406238"} Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.414071 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t98c7" event={"ID":"19abbd38-da74-4e94-94c5-9461701ac88d","Type":"ContainerStarted","Data":"cc2186727486a05084141bd537b8d6233d86270eb410f202c3af4da46618d59c"} Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.416303 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" event={"ID":"54c65dfa-5787-4f1f-979c-9e5daf517650","Type":"ContainerStarted","Data":"abfc89ffa169a3bf567891f19a5722c187642715f1a509b4706adf47c39630a7"} Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.416376 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" event={"ID":"54c65dfa-5787-4f1f-979c-9e5daf517650","Type":"ContainerStarted","Data":"056db4f59ea94ab8e09bedd015c01ec17b0faf85145d15e311d3baec0290f12c"} Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.416975 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.440519 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z5d6l"] Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.442088 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:03 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:03 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:03 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.442162 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.443747 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.446594 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.452010 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z5d6l"] Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.467712 4739 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.472455 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-t98c7" podStartSLOduration=10.472416471 podStartE2EDuration="10.472416471s" podCreationTimestamp="2025-12-01 15:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:03.468658439 +0000 UTC m=+125.294404523" watchObservedRunningTime="2025-12-01 15:37:03.472416471 +0000 UTC m=+125.298162565" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.507905 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" podStartSLOduration=105.507873312 podStartE2EDuration="1m45.507873312s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:03.487689793 +0000 UTC m=+125.313435897" watchObservedRunningTime="2025-12-01 15:37:03.507873312 +0000 UTC m=+125.333619406" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.515141 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbp29\" (UniqueName: \"kubernetes.io/projected/44224079-e66e-4e6f-a74e-5595d3a378f2-kube-api-access-jbp29\") pod \"redhat-marketplace-z5d6l\" (UID: \"44224079-e66e-4e6f-a74e-5595d3a378f2\") " pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.515309 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44224079-e66e-4e6f-a74e-5595d3a378f2-utilities\") pod \"redhat-marketplace-z5d6l\" (UID: \"44224079-e66e-4e6f-a74e-5595d3a378f2\") " pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.515333 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44224079-e66e-4e6f-a74e-5595d3a378f2-catalog-content\") pod \"redhat-marketplace-z5d6l\" (UID: \"44224079-e66e-4e6f-a74e-5595d3a378f2\") " pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.519073 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2bhxd"] Dec 01 15:37:03 crc kubenswrapper[4739]: W1201 15:37:03.595643 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c29aacd_8294_4a12_b340_253e25ebe298.slice/crio-9c41fcbec5dbcaf4c2a7bf7ed78d39fe6b14cf6cbfcbc684a6a7cc61b8a91815 WatchSource:0}: Error finding container 9c41fcbec5dbcaf4c2a7bf7ed78d39fe6b14cf6cbfcbc684a6a7cc61b8a91815: Status 404 returned error can't find the container with id 9c41fcbec5dbcaf4c2a7bf7ed78d39fe6b14cf6cbfcbc684a6a7cc61b8a91815 Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.616494 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44224079-e66e-4e6f-a74e-5595d3a378f2-utilities\") pod \"redhat-marketplace-z5d6l\" (UID: \"44224079-e66e-4e6f-a74e-5595d3a378f2\") " pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.616535 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44224079-e66e-4e6f-a74e-5595d3a378f2-catalog-content\") pod \"redhat-marketplace-z5d6l\" (UID: \"44224079-e66e-4e6f-a74e-5595d3a378f2\") " pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.616612 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbp29\" (UniqueName: \"kubernetes.io/projected/44224079-e66e-4e6f-a74e-5595d3a378f2-kube-api-access-jbp29\") pod \"redhat-marketplace-z5d6l\" (UID: \"44224079-e66e-4e6f-a74e-5595d3a378f2\") " pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.617020 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44224079-e66e-4e6f-a74e-5595d3a378f2-catalog-content\") pod \"redhat-marketplace-z5d6l\" (UID: \"44224079-e66e-4e6f-a74e-5595d3a378f2\") " pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.617123 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44224079-e66e-4e6f-a74e-5595d3a378f2-utilities\") pod \"redhat-marketplace-z5d6l\" (UID: \"44224079-e66e-4e6f-a74e-5595d3a378f2\") " pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.641817 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbp29\" (UniqueName: \"kubernetes.io/projected/44224079-e66e-4e6f-a74e-5595d3a378f2-kube-api-access-jbp29\") pod \"redhat-marketplace-z5d6l\" (UID: \"44224079-e66e-4e6f-a74e-5595d3a378f2\") " pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.692287 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.784962 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.820738 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwdcg\" (UniqueName: \"kubernetes.io/projected/f7e22852-525d-453f-a329-c15a1c5977ad-kube-api-access-kwdcg\") pod \"f7e22852-525d-453f-a329-c15a1c5977ad\" (UID: \"f7e22852-525d-453f-a329-c15a1c5977ad\") " Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.821194 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f7e22852-525d-453f-a329-c15a1c5977ad-config-volume\") pod \"f7e22852-525d-453f-a329-c15a1c5977ad\" (UID: \"f7e22852-525d-453f-a329-c15a1c5977ad\") " Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.821251 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f7e22852-525d-453f-a329-c15a1c5977ad-secret-volume\") pod \"f7e22852-525d-453f-a329-c15a1c5977ad\" (UID: \"f7e22852-525d-453f-a329-c15a1c5977ad\") " Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.821786 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7e22852-525d-453f-a329-c15a1c5977ad-config-volume" (OuterVolumeSpecName: "config-volume") pod "f7e22852-525d-453f-a329-c15a1c5977ad" (UID: "f7e22852-525d-453f-a329-c15a1c5977ad"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.827037 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e22852-525d-453f-a329-c15a1c5977ad-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f7e22852-525d-453f-a329-c15a1c5977ad" (UID: "f7e22852-525d-453f-a329-c15a1c5977ad"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.827255 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l9bd8"] Dec 01 15:37:03 crc kubenswrapper[4739]: E1201 15:37:03.827475 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7e22852-525d-453f-a329-c15a1c5977ad" containerName="collect-profiles" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.827486 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7e22852-525d-453f-a329-c15a1c5977ad" containerName="collect-profiles" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.827593 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7e22852-525d-453f-a329-c15a1c5977ad" containerName="collect-profiles" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.828364 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.835870 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7e22852-525d-453f-a329-c15a1c5977ad-kube-api-access-kwdcg" (OuterVolumeSpecName: "kube-api-access-kwdcg") pod "f7e22852-525d-453f-a329-c15a1c5977ad" (UID: "f7e22852-525d-453f-a329-c15a1c5977ad"). InnerVolumeSpecName "kube-api-access-kwdcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.879035 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l9bd8"] Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.923135 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grgx6\" (UniqueName: \"kubernetes.io/projected/16013e7f-96c5-42ab-ad19-f33b1c0e672c-kube-api-access-grgx6\") pod \"redhat-marketplace-l9bd8\" (UID: \"16013e7f-96c5-42ab-ad19-f33b1c0e672c\") " pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.923182 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16013e7f-96c5-42ab-ad19-f33b1c0e672c-catalog-content\") pod \"redhat-marketplace-l9bd8\" (UID: \"16013e7f-96c5-42ab-ad19-f33b1c0e672c\") " pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.923203 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16013e7f-96c5-42ab-ad19-f33b1c0e672c-utilities\") pod \"redhat-marketplace-l9bd8\" (UID: \"16013e7f-96c5-42ab-ad19-f33b1c0e672c\") " pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.923267 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwdcg\" (UniqueName: \"kubernetes.io/projected/f7e22852-525d-453f-a329-c15a1c5977ad-kube-api-access-kwdcg\") on node \"crc\" DevicePath \"\"" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.923318 4739 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f7e22852-525d-453f-a329-c15a1c5977ad-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 15:37:03 crc kubenswrapper[4739]: I1201 15:37:03.923337 4739 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f7e22852-525d-453f-a329-c15a1c5977ad-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.024388 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grgx6\" (UniqueName: \"kubernetes.io/projected/16013e7f-96c5-42ab-ad19-f33b1c0e672c-kube-api-access-grgx6\") pod \"redhat-marketplace-l9bd8\" (UID: \"16013e7f-96c5-42ab-ad19-f33b1c0e672c\") " pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.024465 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16013e7f-96c5-42ab-ad19-f33b1c0e672c-catalog-content\") pod \"redhat-marketplace-l9bd8\" (UID: \"16013e7f-96c5-42ab-ad19-f33b1c0e672c\") " pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.024494 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16013e7f-96c5-42ab-ad19-f33b1c0e672c-utilities\") pod \"redhat-marketplace-l9bd8\" (UID: \"16013e7f-96c5-42ab-ad19-f33b1c0e672c\") " pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.025400 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16013e7f-96c5-42ab-ad19-f33b1c0e672c-utilities\") pod \"redhat-marketplace-l9bd8\" (UID: \"16013e7f-96c5-42ab-ad19-f33b1c0e672c\") " pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.025947 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16013e7f-96c5-42ab-ad19-f33b1c0e672c-catalog-content\") pod \"redhat-marketplace-l9bd8\" (UID: \"16013e7f-96c5-42ab-ad19-f33b1c0e672c\") " pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.050597 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grgx6\" (UniqueName: \"kubernetes.io/projected/16013e7f-96c5-42ab-ad19-f33b1c0e672c-kube-api-access-grgx6\") pod \"redhat-marketplace-l9bd8\" (UID: \"16013e7f-96c5-42ab-ad19-f33b1c0e672c\") " pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.063182 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z5d6l"] Dec 01 15:37:04 crc kubenswrapper[4739]: W1201 15:37:04.070561 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44224079_e66e_4e6f_a74e_5595d3a378f2.slice/crio-b9ef385aeb24babf07f34492de4a061dfc56e3613a4b034539d5203a3e807332 WatchSource:0}: Error finding container b9ef385aeb24babf07f34492de4a061dfc56e3613a4b034539d5203a3e807332: Status 404 returned error can't find the container with id b9ef385aeb24babf07f34492de4a061dfc56e3613a4b034539d5203a3e807332 Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.168513 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.369315 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l9bd8"] Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.424718 4739 generic.go:334] "Generic (PLEG): container finished" podID="4c29aacd-8294-4a12-b340-253e25ebe298" containerID="13732b0fc46b3e3efce48e313887e8adfb880a9a365daff07a6152e09abcafa9" exitCode=0 Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.424823 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2bhxd" event={"ID":"4c29aacd-8294-4a12-b340-253e25ebe298","Type":"ContainerDied","Data":"13732b0fc46b3e3efce48e313887e8adfb880a9a365daff07a6152e09abcafa9"} Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.424907 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2bhxd" event={"ID":"4c29aacd-8294-4a12-b340-253e25ebe298","Type":"ContainerStarted","Data":"9c41fcbec5dbcaf4c2a7bf7ed78d39fe6b14cf6cbfcbc684a6a7cc61b8a91815"} Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.428054 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"be4aaf21-24c7-4500-8151-07af3fab2bcb","Type":"ContainerStarted","Data":"000c1afb32a42f69653c5b67678ca6309478551de12b957b0ee0dac7cb33df3b"} Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.428121 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"be4aaf21-24c7-4500-8151-07af3fab2bcb","Type":"ContainerStarted","Data":"9964696ec155fae44ba4912e2900fc8d44e78b90ab982c7f54fc12c7685a36f5"} Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.433240 4739 generic.go:334] "Generic (PLEG): container finished" podID="44224079-e66e-4e6f-a74e-5595d3a378f2" containerID="47194f78ba84646b6ffdb103252765d9b384c1ffd85546d88d3d6651ce38b357" exitCode=0 Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.433314 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5d6l" event={"ID":"44224079-e66e-4e6f-a74e-5595d3a378f2","Type":"ContainerDied","Data":"47194f78ba84646b6ffdb103252765d9b384c1ffd85546d88d3d6651ce38b357"} Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.433347 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5d6l" event={"ID":"44224079-e66e-4e6f-a74e-5595d3a378f2","Type":"ContainerStarted","Data":"b9ef385aeb24babf07f34492de4a061dfc56e3613a4b034539d5203a3e807332"} Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.440451 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" event={"ID":"f7e22852-525d-453f-a329-c15a1c5977ad","Type":"ContainerDied","Data":"6cd5ff7693a3e2323b95972f82d4ec46c9b5381541c988c4dbaf6021d4d2749e"} Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.440489 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cd5ff7693a3e2323b95972f82d4ec46c9b5381541c988c4dbaf6021d4d2749e" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.440493 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.447576 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:04 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:04 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:04 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.447650 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.450698 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9bd8" event={"ID":"16013e7f-96c5-42ab-ad19-f33b1c0e672c","Type":"ContainerStarted","Data":"bff74b14c9462f5aee08189d4b39fbbdf2f898dba75073493cd189239f966cca"} Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.467086 4739 generic.go:334] "Generic (PLEG): container finished" podID="68008673-38d1-477e-8963-1f4d069ba0c7" containerID="1e5a101dc84bb98fde41bc27c306513659c4dd027bba302f99733d9728a80dd3" exitCode=0 Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.467452 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2t8q" event={"ID":"68008673-38d1-477e-8963-1f4d069ba0c7","Type":"ContainerDied","Data":"1e5a101dc84bb98fde41bc27c306513659c4dd027bba302f99733d9728a80dd3"} Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.495668 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.49564239 podStartE2EDuration="2.49564239s" podCreationTimestamp="2025-12-01 15:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:04.494093054 +0000 UTC m=+126.319839158" watchObservedRunningTime="2025-12-01 15:37:04.49564239 +0000 UTC m=+126.321388494" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.518662 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.601606 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.605479 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.609103 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.610308 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.610504 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.638336 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/59f920e1-858b-46db-a50a-d42f758b3fe2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"59f920e1-858b-46db-a50a-d42f758b3fe2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.638500 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/59f920e1-858b-46db-a50a-d42f758b3fe2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"59f920e1-858b-46db-a50a-d42f758b3fe2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.739929 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/59f920e1-858b-46db-a50a-d42f758b3fe2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"59f920e1-858b-46db-a50a-d42f758b3fe2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.740045 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/59f920e1-858b-46db-a50a-d42f758b3fe2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"59f920e1-858b-46db-a50a-d42f758b3fe2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.740355 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/59f920e1-858b-46db-a50a-d42f758b3fe2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"59f920e1-858b-46db-a50a-d42f758b3fe2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.760007 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/59f920e1-858b-46db-a50a-d42f758b3fe2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"59f920e1-858b-46db-a50a-d42f758b3fe2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.830186 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ksr6g"] Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.831599 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.834102 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.836812 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ksr6g"] Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.932214 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.944565 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4s5b\" (UniqueName: \"kubernetes.io/projected/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-kube-api-access-c4s5b\") pod \"redhat-operators-ksr6g\" (UID: \"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8\") " pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.944691 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-utilities\") pod \"redhat-operators-ksr6g\" (UID: \"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8\") " pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:37:04 crc kubenswrapper[4739]: I1201 15:37:04.944730 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-catalog-content\") pod \"redhat-operators-ksr6g\" (UID: \"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8\") " pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.046157 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-utilities\") pod \"redhat-operators-ksr6g\" (UID: \"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8\") " pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.046224 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-catalog-content\") pod \"redhat-operators-ksr6g\" (UID: \"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8\") " pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.046316 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4s5b\" (UniqueName: \"kubernetes.io/projected/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-kube-api-access-c4s5b\") pod \"redhat-operators-ksr6g\" (UID: \"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8\") " pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.047143 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-utilities\") pod \"redhat-operators-ksr6g\" (UID: \"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8\") " pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.047353 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-catalog-content\") pod \"redhat-operators-ksr6g\" (UID: \"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8\") " pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.078233 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4s5b\" (UniqueName: \"kubernetes.io/projected/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-kube-api-access-c4s5b\") pod \"redhat-operators-ksr6g\" (UID: \"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8\") " pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.111399 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.150677 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.222616 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xwdtf"] Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.224333 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.234301 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xwdtf"] Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.346068 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ksr6g"] Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.349890 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfbd2a39-0be3-48df-8797-55b4e72603d7-utilities\") pod \"redhat-operators-xwdtf\" (UID: \"bfbd2a39-0be3-48df-8797-55b4e72603d7\") " pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.350047 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbz2g\" (UniqueName: \"kubernetes.io/projected/bfbd2a39-0be3-48df-8797-55b4e72603d7-kube-api-access-wbz2g\") pod \"redhat-operators-xwdtf\" (UID: \"bfbd2a39-0be3-48df-8797-55b4e72603d7\") " pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.350098 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfbd2a39-0be3-48df-8797-55b4e72603d7-catalog-content\") pod \"redhat-operators-xwdtf\" (UID: \"bfbd2a39-0be3-48df-8797-55b4e72603d7\") " pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:05 crc kubenswrapper[4739]: W1201 15:37:05.363049 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83a6fb61_1d84_43a3_bc8e_5890c3c0a4f8.slice/crio-e09b33731de92ceffd80e1b762d140bf5cb8ff1c555c6d7712eed6bfb930e708 WatchSource:0}: Error finding container e09b33731de92ceffd80e1b762d140bf5cb8ff1c555c6d7712eed6bfb930e708: Status 404 returned error can't find the container with id e09b33731de92ceffd80e1b762d140bf5cb8ff1c555c6d7712eed6bfb930e708 Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.402562 4739 patch_prober.go:28] interesting pod/downloads-7954f5f757-s6nss container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.402600 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-s6nss" podUID="67608617-45fd-4c45-87f9-325996855e5e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.402939 4739 patch_prober.go:28] interesting pod/downloads-7954f5f757-s6nss container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.402958 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-s6nss" podUID="67608617-45fd-4c45-87f9-325996855e5e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.435436 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.438256 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:05 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:05 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:05 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.438320 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.451616 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfbd2a39-0be3-48df-8797-55b4e72603d7-catalog-content\") pod \"redhat-operators-xwdtf\" (UID: \"bfbd2a39-0be3-48df-8797-55b4e72603d7\") " pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.451663 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfbd2a39-0be3-48df-8797-55b4e72603d7-utilities\") pod \"redhat-operators-xwdtf\" (UID: \"bfbd2a39-0be3-48df-8797-55b4e72603d7\") " pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.451773 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbz2g\" (UniqueName: \"kubernetes.io/projected/bfbd2a39-0be3-48df-8797-55b4e72603d7-kube-api-access-wbz2g\") pod \"redhat-operators-xwdtf\" (UID: \"bfbd2a39-0be3-48df-8797-55b4e72603d7\") " pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.452099 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfbd2a39-0be3-48df-8797-55b4e72603d7-catalog-content\") pod \"redhat-operators-xwdtf\" (UID: \"bfbd2a39-0be3-48df-8797-55b4e72603d7\") " pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.452295 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfbd2a39-0be3-48df-8797-55b4e72603d7-utilities\") pod \"redhat-operators-xwdtf\" (UID: \"bfbd2a39-0be3-48df-8797-55b4e72603d7\") " pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.469062 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbz2g\" (UniqueName: \"kubernetes.io/projected/bfbd2a39-0be3-48df-8797-55b4e72603d7-kube-api-access-wbz2g\") pod \"redhat-operators-xwdtf\" (UID: \"bfbd2a39-0be3-48df-8797-55b4e72603d7\") " pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.479598 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ksr6g" event={"ID":"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8","Type":"ContainerStarted","Data":"e09b33731de92ceffd80e1b762d140bf5cb8ff1c555c6d7712eed6bfb930e708"} Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.482820 4739 generic.go:334] "Generic (PLEG): container finished" podID="16013e7f-96c5-42ab-ad19-f33b1c0e672c" containerID="943787895eda2bfcbe4ed1faccdda17dad595de3d8ccb0c68ae8cf215935e2c7" exitCode=0 Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.482882 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9bd8" event={"ID":"16013e7f-96c5-42ab-ad19-f33b1c0e672c","Type":"ContainerDied","Data":"943787895eda2bfcbe4ed1faccdda17dad595de3d8ccb0c68ae8cf215935e2c7"} Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.488650 4739 generic.go:334] "Generic (PLEG): container finished" podID="be4aaf21-24c7-4500-8151-07af3fab2bcb" containerID="000c1afb32a42f69653c5b67678ca6309478551de12b957b0ee0dac7cb33df3b" exitCode=0 Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.488734 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"be4aaf21-24c7-4500-8151-07af3fab2bcb","Type":"ContainerDied","Data":"000c1afb32a42f69653c5b67678ca6309478551de12b957b0ee0dac7cb33df3b"} Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.492558 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"59f920e1-858b-46db-a50a-d42f758b3fe2","Type":"ContainerStarted","Data":"663da1eb02e10c9b410693ac01b86bc1e8cf01e9fb9a90d0a446cf0e683dee21"} Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.501605 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.503929 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.513312 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.555153 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.621483 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.621530 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.628022 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.710088 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.711424 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.712976 4739 patch_prober.go:28] interesting pod/console-f9d7485db-5dxtl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.713047 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5dxtl" podUID="64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" containerName="console" probeResult="failure" output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" Dec 01 15:37:05 crc kubenswrapper[4739]: I1201 15:37:05.910772 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xwdtf"] Dec 01 15:37:05 crc kubenswrapper[4739]: W1201 15:37:05.970864 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfbd2a39_0be3_48df_8797_55b4e72603d7.slice/crio-6e40d5603e47bdb95b304594728672ddc5f2c4d8ea640a86126921a67f984b2d WatchSource:0}: Error finding container 6e40d5603e47bdb95b304594728672ddc5f2c4d8ea640a86126921a67f984b2d: Status 404 returned error can't find the container with id 6e40d5603e47bdb95b304594728672ddc5f2c4d8ea640a86126921a67f984b2d Dec 01 15:37:06 crc kubenswrapper[4739]: I1201 15:37:06.439183 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:06 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:06 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:06 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:06 crc kubenswrapper[4739]: I1201 15:37:06.439653 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:06 crc kubenswrapper[4739]: I1201 15:37:06.512637 4739 generic.go:334] "Generic (PLEG): container finished" podID="59f920e1-858b-46db-a50a-d42f758b3fe2" containerID="e9b16560078ad2d23681f659c25156e5906cf8098b0a7618411a672a6fcc8a3b" exitCode=0 Dec 01 15:37:06 crc kubenswrapper[4739]: I1201 15:37:06.512721 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"59f920e1-858b-46db-a50a-d42f758b3fe2","Type":"ContainerDied","Data":"e9b16560078ad2d23681f659c25156e5906cf8098b0a7618411a672a6fcc8a3b"} Dec 01 15:37:06 crc kubenswrapper[4739]: I1201 15:37:06.516191 4739 generic.go:334] "Generic (PLEG): container finished" podID="bfbd2a39-0be3-48df-8797-55b4e72603d7" containerID="8b20bbc5f49dd9b0c6990e0a0e7a8c0bba83d323f9bfdebcc08405b29ffc3d7f" exitCode=0 Dec 01 15:37:06 crc kubenswrapper[4739]: I1201 15:37:06.516251 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwdtf" event={"ID":"bfbd2a39-0be3-48df-8797-55b4e72603d7","Type":"ContainerDied","Data":"8b20bbc5f49dd9b0c6990e0a0e7a8c0bba83d323f9bfdebcc08405b29ffc3d7f"} Dec 01 15:37:06 crc kubenswrapper[4739]: I1201 15:37:06.516273 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwdtf" event={"ID":"bfbd2a39-0be3-48df-8797-55b4e72603d7","Type":"ContainerStarted","Data":"6e40d5603e47bdb95b304594728672ddc5f2c4d8ea640a86126921a67f984b2d"} Dec 01 15:37:06 crc kubenswrapper[4739]: I1201 15:37:06.520851 4739 generic.go:334] "Generic (PLEG): container finished" podID="83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8" containerID="e4f1ae38209d3755d683eb360ba83057d74322add4086b2c1c6b962a5d8d37c8" exitCode=0 Dec 01 15:37:06 crc kubenswrapper[4739]: I1201 15:37:06.521046 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ksr6g" event={"ID":"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8","Type":"ContainerDied","Data":"e4f1ae38209d3755d683eb360ba83057d74322add4086b2c1c6b962a5d8d37c8"} Dec 01 15:37:06 crc kubenswrapper[4739]: I1201 15:37:06.536629 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-w9l4v" Dec 01 15:37:06 crc kubenswrapper[4739]: I1201 15:37:06.542881 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8bq5d" Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.098694 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.188992 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be4aaf21-24c7-4500-8151-07af3fab2bcb-kube-api-access\") pod \"be4aaf21-24c7-4500-8151-07af3fab2bcb\" (UID: \"be4aaf21-24c7-4500-8151-07af3fab2bcb\") " Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.189187 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be4aaf21-24c7-4500-8151-07af3fab2bcb-kubelet-dir\") pod \"be4aaf21-24c7-4500-8151-07af3fab2bcb\" (UID: \"be4aaf21-24c7-4500-8151-07af3fab2bcb\") " Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.189772 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be4aaf21-24c7-4500-8151-07af3fab2bcb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "be4aaf21-24c7-4500-8151-07af3fab2bcb" (UID: "be4aaf21-24c7-4500-8151-07af3fab2bcb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.190011 4739 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be4aaf21-24c7-4500-8151-07af3fab2bcb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.196181 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be4aaf21-24c7-4500-8151-07af3fab2bcb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "be4aaf21-24c7-4500-8151-07af3fab2bcb" (UID: "be4aaf21-24c7-4500-8151-07af3fab2bcb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.291147 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be4aaf21-24c7-4500-8151-07af3fab2bcb-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.439569 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:07 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:07 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:07 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.439639 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.538123 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.538132 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"be4aaf21-24c7-4500-8151-07af3fab2bcb","Type":"ContainerDied","Data":"9964696ec155fae44ba4912e2900fc8d44e78b90ab982c7f54fc12c7685a36f5"} Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.538176 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9964696ec155fae44ba4912e2900fc8d44e78b90ab982c7f54fc12c7685a36f5" Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.791974 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.909030 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/59f920e1-858b-46db-a50a-d42f758b3fe2-kubelet-dir\") pod \"59f920e1-858b-46db-a50a-d42f758b3fe2\" (UID: \"59f920e1-858b-46db-a50a-d42f758b3fe2\") " Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.909116 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/59f920e1-858b-46db-a50a-d42f758b3fe2-kube-api-access\") pod \"59f920e1-858b-46db-a50a-d42f758b3fe2\" (UID: \"59f920e1-858b-46db-a50a-d42f758b3fe2\") " Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.909340 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/59f920e1-858b-46db-a50a-d42f758b3fe2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "59f920e1-858b-46db-a50a-d42f758b3fe2" (UID: "59f920e1-858b-46db-a50a-d42f758b3fe2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.909901 4739 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/59f920e1-858b-46db-a50a-d42f758b3fe2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 15:37:07 crc kubenswrapper[4739]: I1201 15:37:07.912455 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59f920e1-858b-46db-a50a-d42f758b3fe2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "59f920e1-858b-46db-a50a-d42f758b3fe2" (UID: "59f920e1-858b-46db-a50a-d42f758b3fe2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:37:08 crc kubenswrapper[4739]: I1201 15:37:08.011145 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/59f920e1-858b-46db-a50a-d42f758b3fe2-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 15:37:08 crc kubenswrapper[4739]: I1201 15:37:08.444995 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:08 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:08 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:08 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:08 crc kubenswrapper[4739]: I1201 15:37:08.445081 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:08 crc kubenswrapper[4739]: I1201 15:37:08.550833 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"59f920e1-858b-46db-a50a-d42f758b3fe2","Type":"ContainerDied","Data":"663da1eb02e10c9b410693ac01b86bc1e8cf01e9fb9a90d0a446cf0e683dee21"} Dec 01 15:37:08 crc kubenswrapper[4739]: I1201 15:37:08.550855 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 15:37:08 crc kubenswrapper[4739]: I1201 15:37:08.550887 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="663da1eb02e10c9b410693ac01b86bc1e8cf01e9fb9a90d0a446cf0e683dee21" Dec 01 15:37:09 crc kubenswrapper[4739]: I1201 15:37:09.440691 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:09 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:09 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:09 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:09 crc kubenswrapper[4739]: I1201 15:37:09.440776 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:10 crc kubenswrapper[4739]: I1201 15:37:10.438238 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:10 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:10 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:10 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:10 crc kubenswrapper[4739]: I1201 15:37:10.438637 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:11 crc kubenswrapper[4739]: I1201 15:37:11.307690 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-99fnd" Dec 01 15:37:11 crc kubenswrapper[4739]: I1201 15:37:11.437886 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:11 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:11 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:11 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:11 crc kubenswrapper[4739]: I1201 15:37:11.438077 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:12 crc kubenswrapper[4739]: I1201 15:37:12.437966 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:12 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:12 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:12 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:12 crc kubenswrapper[4739]: I1201 15:37:12.438022 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:13 crc kubenswrapper[4739]: I1201 15:37:13.441368 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:13 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:13 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:13 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:13 crc kubenswrapper[4739]: I1201 15:37:13.441824 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:14 crc kubenswrapper[4739]: I1201 15:37:14.438494 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:14 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:14 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:14 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:14 crc kubenswrapper[4739]: I1201 15:37:14.438607 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:15 crc kubenswrapper[4739]: I1201 15:37:15.418459 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-s6nss" Dec 01 15:37:15 crc kubenswrapper[4739]: I1201 15:37:15.439059 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:15 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:15 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:15 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:15 crc kubenswrapper[4739]: I1201 15:37:15.439138 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:15 crc kubenswrapper[4739]: I1201 15:37:15.710698 4739 patch_prober.go:28] interesting pod/console-f9d7485db-5dxtl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Dec 01 15:37:15 crc kubenswrapper[4739]: I1201 15:37:15.710786 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5dxtl" podUID="64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" containerName="console" probeResult="failure" output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" Dec 01 15:37:16 crc kubenswrapper[4739]: I1201 15:37:16.438363 4739 patch_prober.go:28] interesting pod/router-default-5444994796-6b8xv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 15:37:16 crc kubenswrapper[4739]: [-]has-synced failed: reason withheld Dec 01 15:37:16 crc kubenswrapper[4739]: [+]process-running ok Dec 01 15:37:16 crc kubenswrapper[4739]: healthz check failed Dec 01 15:37:16 crc kubenswrapper[4739]: I1201 15:37:16.438850 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6b8xv" podUID="217d2158-0c6e-48d9-9713-5499a54655d9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 15:37:17 crc kubenswrapper[4739]: I1201 15:37:17.438397 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:37:17 crc kubenswrapper[4739]: I1201 15:37:17.442462 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-6b8xv" Dec 01 15:37:22 crc kubenswrapper[4739]: I1201 15:37:22.446229 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:37:24 crc kubenswrapper[4739]: I1201 15:37:24.716682 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:37:24 crc kubenswrapper[4739]: I1201 15:37:24.719437 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 15:37:24 crc kubenswrapper[4739]: I1201 15:37:24.743771 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:37:24 crc kubenswrapper[4739]: I1201 15:37:24.819721 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:37:24 crc kubenswrapper[4739]: I1201 15:37:24.819812 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:37:24 crc kubenswrapper[4739]: I1201 15:37:24.819860 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:37:24 crc kubenswrapper[4739]: I1201 15:37:24.822960 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 15:37:24 crc kubenswrapper[4739]: I1201 15:37:24.822961 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 15:37:24 crc kubenswrapper[4739]: I1201 15:37:24.833190 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 15:37:24 crc kubenswrapper[4739]: I1201 15:37:24.844042 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:37:24 crc kubenswrapper[4739]: I1201 15:37:24.844179 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:37:24 crc kubenswrapper[4739]: I1201 15:37:24.860114 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:37:25 crc kubenswrapper[4739]: I1201 15:37:25.100072 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:37:25 crc kubenswrapper[4739]: I1201 15:37:25.109374 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 15:37:25 crc kubenswrapper[4739]: I1201 15:37:25.115828 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 15:37:25 crc kubenswrapper[4739]: E1201 15:37:25.285002 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 01 15:37:25 crc kubenswrapper[4739]: E1201 15:37:25.285314 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-75fd8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-8vzjc_openshift-marketplace(cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 15:37:25 crc kubenswrapper[4739]: E1201 15:37:25.287660 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-8vzjc" podUID="cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" Dec 01 15:37:25 crc kubenswrapper[4739]: I1201 15:37:25.720111 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:37:25 crc kubenswrapper[4739]: I1201 15:37:25.730347 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:37:30 crc kubenswrapper[4739]: E1201 15:37:30.005603 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-8vzjc" podUID="cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" Dec 01 15:37:30 crc kubenswrapper[4739]: E1201 15:37:30.064752 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 01 15:37:30 crc kubenswrapper[4739]: E1201 15:37:30.064942 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dpjkw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-r2t8q_openshift-marketplace(68008673-38d1-477e-8963-1f4d069ba0c7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 15:37:30 crc kubenswrapper[4739]: E1201 15:37:30.066554 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-r2t8q" podUID="68008673-38d1-477e-8963-1f4d069ba0c7" Dec 01 15:37:31 crc kubenswrapper[4739]: E1201 15:37:31.421458 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-r2t8q" podUID="68008673-38d1-477e-8963-1f4d069ba0c7" Dec 01 15:37:31 crc kubenswrapper[4739]: E1201 15:37:31.661647 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 01 15:37:31 crc kubenswrapper[4739]: E1201 15:37:31.661878 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-grgx6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-l9bd8_openshift-marketplace(16013e7f-96c5-42ab-ad19-f33b1c0e672c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 15:37:31 crc kubenswrapper[4739]: E1201 15:37:31.663035 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-l9bd8" podUID="16013e7f-96c5-42ab-ad19-f33b1c0e672c" Dec 01 15:37:32 crc kubenswrapper[4739]: E1201 15:37:32.014128 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 01 15:37:32 crc kubenswrapper[4739]: E1201 15:37:32.014351 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zlwr9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-nwqtt_openshift-marketplace(5b9930dd-3336-438d-84fc-3a62b6a86e75): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 15:37:32 crc kubenswrapper[4739]: E1201 15:37:32.015833 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-nwqtt" podUID="5b9930dd-3336-438d-84fc-3a62b6a86e75" Dec 01 15:37:32 crc kubenswrapper[4739]: E1201 15:37:32.333485 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 01 15:37:32 crc kubenswrapper[4739]: E1201 15:37:32.333688 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jbp29,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-z5d6l_openshift-marketplace(44224079-e66e-4e6f-a74e-5595d3a378f2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 15:37:32 crc kubenswrapper[4739]: E1201 15:37:32.336040 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-z5d6l" podUID="44224079-e66e-4e6f-a74e-5595d3a378f2" Dec 01 15:37:32 crc kubenswrapper[4739]: E1201 15:37:32.423105 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 01 15:37:32 crc kubenswrapper[4739]: E1201 15:37:32.423528 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k5dkv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-2bhxd_openshift-marketplace(4c29aacd-8294-4a12-b340-253e25ebe298): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 15:37:32 crc kubenswrapper[4739]: E1201 15:37:32.425566 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-2bhxd" podUID="4c29aacd-8294-4a12-b340-253e25ebe298" Dec 01 15:37:34 crc kubenswrapper[4739]: E1201 15:37:34.552289 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-nwqtt" podUID="5b9930dd-3336-438d-84fc-3a62b6a86e75" Dec 01 15:37:34 crc kubenswrapper[4739]: E1201 15:37:34.552299 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-l9bd8" podUID="16013e7f-96c5-42ab-ad19-f33b1c0e672c" Dec 01 15:37:34 crc kubenswrapper[4739]: E1201 15:37:34.552411 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-2bhxd" podUID="4c29aacd-8294-4a12-b340-253e25ebe298" Dec 01 15:37:34 crc kubenswrapper[4739]: E1201 15:37:34.553139 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-z5d6l" podUID="44224079-e66e-4e6f-a74e-5595d3a378f2" Dec 01 15:37:35 crc kubenswrapper[4739]: I1201 15:37:35.752749 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4f917d8fc1a61bc6252320d3fb8b1d3479b7e64f8188834a2aedf81a978fafb4"} Dec 01 15:37:35 crc kubenswrapper[4739]: I1201 15:37:35.753461 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"74ce3df72de92fe778c477bcf7797fd0dc84c683824a2c29b90b49b9c597eb86"} Dec 01 15:37:35 crc kubenswrapper[4739]: I1201 15:37:35.754943 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"8ba2bd15af8dffbc71790e71def8ad43124b3c8059e8b58a9518c4146547ebd0"} Dec 01 15:37:35 crc kubenswrapper[4739]: I1201 15:37:35.754973 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"5627a2afcb4df851a08c89b5bff1c7cb0106425dbdb6bb6ddf9d578c5029d531"} Dec 01 15:37:35 crc kubenswrapper[4739]: I1201 15:37:35.757379 4739 generic.go:334] "Generic (PLEG): container finished" podID="bfbd2a39-0be3-48df-8797-55b4e72603d7" containerID="690f591b30eda2517b85c7ddcd303f3816a2bd6f20840dc26c61ff274a8504ad" exitCode=0 Dec 01 15:37:35 crc kubenswrapper[4739]: I1201 15:37:35.757444 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwdtf" event={"ID":"bfbd2a39-0be3-48df-8797-55b4e72603d7","Type":"ContainerDied","Data":"690f591b30eda2517b85c7ddcd303f3816a2bd6f20840dc26c61ff274a8504ad"} Dec 01 15:37:35 crc kubenswrapper[4739]: I1201 15:37:35.759654 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"82d0331dbc6d9cf6e822bb9abd628d341ce0da0dd9da1819d01ddecfb6ad4333"} Dec 01 15:37:35 crc kubenswrapper[4739]: I1201 15:37:35.759697 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"831dbe7bf0511b0c167933f08bbc20cce070224782a6684048f3c04ade59bfae"} Dec 01 15:37:35 crc kubenswrapper[4739]: I1201 15:37:35.760429 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:37:35 crc kubenswrapper[4739]: I1201 15:37:35.763036 4739 generic.go:334] "Generic (PLEG): container finished" podID="83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8" containerID="7b700d8f868530db267924bb4d5e23c79742cc9babc6ef9b86a223eec950efb4" exitCode=0 Dec 01 15:37:35 crc kubenswrapper[4739]: I1201 15:37:35.763089 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ksr6g" event={"ID":"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8","Type":"ContainerDied","Data":"7b700d8f868530db267924bb4d5e23c79742cc9babc6ef9b86a223eec950efb4"} Dec 01 15:37:36 crc kubenswrapper[4739]: I1201 15:37:36.233566 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2c4qm" Dec 01 15:37:36 crc kubenswrapper[4739]: I1201 15:37:36.770408 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ksr6g" event={"ID":"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8","Type":"ContainerStarted","Data":"ff70135236a4e596926d23d394ea708c66ab95f21964efd0cb13000422d59f74"} Dec 01 15:37:36 crc kubenswrapper[4739]: I1201 15:37:36.772815 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwdtf" event={"ID":"bfbd2a39-0be3-48df-8797-55b4e72603d7","Type":"ContainerStarted","Data":"79eee8a23eb277a3572615023fd650c2e620e91a63e18bfeea078a2d2b855155"} Dec 01 15:37:36 crc kubenswrapper[4739]: I1201 15:37:36.789597 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ksr6g" podStartSLOduration=3.110888325 podStartE2EDuration="32.789573965s" podCreationTimestamp="2025-12-01 15:37:04 +0000 UTC" firstStartedPulling="2025-12-01 15:37:06.522981031 +0000 UTC m=+128.348727125" lastFinishedPulling="2025-12-01 15:37:36.201666631 +0000 UTC m=+158.027412765" observedRunningTime="2025-12-01 15:37:36.78746568 +0000 UTC m=+158.613211794" watchObservedRunningTime="2025-12-01 15:37:36.789573965 +0000 UTC m=+158.615320059" Dec 01 15:37:36 crc kubenswrapper[4739]: I1201 15:37:36.806503 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xwdtf" podStartSLOduration=2.034387594 podStartE2EDuration="31.806481568s" podCreationTimestamp="2025-12-01 15:37:05 +0000 UTC" firstStartedPulling="2025-12-01 15:37:06.517796748 +0000 UTC m=+128.343542842" lastFinishedPulling="2025-12-01 15:37:36.289890722 +0000 UTC m=+158.115636816" observedRunningTime="2025-12-01 15:37:36.804963882 +0000 UTC m=+158.630709996" watchObservedRunningTime="2025-12-01 15:37:36.806481568 +0000 UTC m=+158.632227682" Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.505629 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 15:37:37 crc kubenswrapper[4739]: E1201 15:37:37.508999 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be4aaf21-24c7-4500-8151-07af3fab2bcb" containerName="pruner" Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.509057 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="be4aaf21-24c7-4500-8151-07af3fab2bcb" containerName="pruner" Dec 01 15:37:37 crc kubenswrapper[4739]: E1201 15:37:37.509082 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f920e1-858b-46db-a50a-d42f758b3fe2" containerName="pruner" Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.509091 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f920e1-858b-46db-a50a-d42f758b3fe2" containerName="pruner" Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.509199 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="59f920e1-858b-46db-a50a-d42f758b3fe2" containerName="pruner" Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.509216 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="be4aaf21-24c7-4500-8151-07af3fab2bcb" containerName="pruner" Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.509757 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.511596 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.512659 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.519147 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.637954 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81451049-3d54-44c9-9c89-fd336f199313-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"81451049-3d54-44c9-9c89-fd336f199313\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.638109 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/81451049-3d54-44c9-9c89-fd336f199313-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"81451049-3d54-44c9-9c89-fd336f199313\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.739580 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/81451049-3d54-44c9-9c89-fd336f199313-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"81451049-3d54-44c9-9c89-fd336f199313\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.739689 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81451049-3d54-44c9-9c89-fd336f199313-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"81451049-3d54-44c9-9c89-fd336f199313\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.740591 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/81451049-3d54-44c9-9c89-fd336f199313-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"81451049-3d54-44c9-9c89-fd336f199313\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.763460 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81451049-3d54-44c9-9c89-fd336f199313-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"81451049-3d54-44c9-9c89-fd336f199313\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 15:37:37 crc kubenswrapper[4739]: I1201 15:37:37.840543 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 15:37:38 crc kubenswrapper[4739]: I1201 15:37:38.283377 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 15:37:38 crc kubenswrapper[4739]: I1201 15:37:38.785326 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"81451049-3d54-44c9-9c89-fd336f199313","Type":"ContainerStarted","Data":"0ac6d328d2953440c5efd3412616d0cd2e9119cb8ca3c398dae7d2bc8d9ae28f"} Dec 01 15:37:38 crc kubenswrapper[4739]: I1201 15:37:38.785701 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"81451049-3d54-44c9-9c89-fd336f199313","Type":"ContainerStarted","Data":"3c0baa3409c7580a0c6eb5a1a3ec55733e933bf41be60cdbb8d9afc2f727cb23"} Dec 01 15:37:39 crc kubenswrapper[4739]: I1201 15:37:39.622547 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:37:39 crc kubenswrapper[4739]: I1201 15:37:39.622612 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:37:39 crc kubenswrapper[4739]: I1201 15:37:39.793473 4739 generic.go:334] "Generic (PLEG): container finished" podID="81451049-3d54-44c9-9c89-fd336f199313" containerID="0ac6d328d2953440c5efd3412616d0cd2e9119cb8ca3c398dae7d2bc8d9ae28f" exitCode=0 Dec 01 15:37:39 crc kubenswrapper[4739]: I1201 15:37:39.793593 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"81451049-3d54-44c9-9c89-fd336f199313","Type":"ContainerDied","Data":"0ac6d328d2953440c5efd3412616d0cd2e9119cb8ca3c398dae7d2bc8d9ae28f"} Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.036884 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.094556 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs\") pod \"network-metrics-daemon-bsjx4\" (UID: \"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\") " pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.096758 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.112751 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55ed5d4d-1f46-4e38-9176-2d1a8e5417cf-metrics-certs\") pod \"network-metrics-daemon-bsjx4\" (UID: \"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf\") " pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.195704 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/81451049-3d54-44c9-9c89-fd336f199313-kubelet-dir\") pod \"81451049-3d54-44c9-9c89-fd336f199313\" (UID: \"81451049-3d54-44c9-9c89-fd336f199313\") " Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.195781 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81451049-3d54-44c9-9c89-fd336f199313-kube-api-access\") pod \"81451049-3d54-44c9-9c89-fd336f199313\" (UID: \"81451049-3d54-44c9-9c89-fd336f199313\") " Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.195869 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81451049-3d54-44c9-9c89-fd336f199313-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "81451049-3d54-44c9-9c89-fd336f199313" (UID: "81451049-3d54-44c9-9c89-fd336f199313"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.196180 4739 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/81451049-3d54-44c9-9c89-fd336f199313-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.202629 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81451049-3d54-44c9-9c89-fd336f199313-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "81451049-3d54-44c9-9c89-fd336f199313" (UID: "81451049-3d54-44c9-9c89-fd336f199313"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.297164 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81451049-3d54-44c9-9c89-fd336f199313-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.329047 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.337662 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bsjx4" Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.583234 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-bsjx4"] Dec 01 15:37:41 crc kubenswrapper[4739]: W1201 15:37:41.592551 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55ed5d4d_1f46_4e38_9176_2d1a8e5417cf.slice/crio-faa0e131e7537f39de7919402025922d2dc1729aa9beb5b2c2f65a1024428cb7 WatchSource:0}: Error finding container faa0e131e7537f39de7919402025922d2dc1729aa9beb5b2c2f65a1024428cb7: Status 404 returned error can't find the container with id faa0e131e7537f39de7919402025922d2dc1729aa9beb5b2c2f65a1024428cb7 Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.809175 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" event={"ID":"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf","Type":"ContainerStarted","Data":"faa0e131e7537f39de7919402025922d2dc1729aa9beb5b2c2f65a1024428cb7"} Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.811136 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"81451049-3d54-44c9-9c89-fd336f199313","Type":"ContainerDied","Data":"3c0baa3409c7580a0c6eb5a1a3ec55733e933bf41be60cdbb8d9afc2f727cb23"} Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.811205 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c0baa3409c7580a0c6eb5a1a3ec55733e933bf41be60cdbb8d9afc2f727cb23" Dec 01 15:37:41 crc kubenswrapper[4739]: I1201 15:37:41.811225 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 15:37:43 crc kubenswrapper[4739]: I1201 15:37:43.823078 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" event={"ID":"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf","Type":"ContainerStarted","Data":"8820813d1d9121665d8f35a1c1430a758c48be4478b54d856ac9e73ce4209469"} Dec 01 15:37:43 crc kubenswrapper[4739]: I1201 15:37:43.823490 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-bsjx4" event={"ID":"55ed5d4d-1f46-4e38-9176-2d1a8e5417cf","Type":"ContainerStarted","Data":"2651d42d67ef8481047536917f40770a2416a189af57995bc5f60c01218d84b3"} Dec 01 15:37:43 crc kubenswrapper[4739]: I1201 15:37:43.845471 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-bsjx4" podStartSLOduration=145.845443077 podStartE2EDuration="2m25.845443077s" podCreationTimestamp="2025-12-01 15:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:43.838945515 +0000 UTC m=+165.664691609" watchObservedRunningTime="2025-12-01 15:37:43.845443077 +0000 UTC m=+165.671189171" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.301742 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 15:37:44 crc kubenswrapper[4739]: E1201 15:37:44.302123 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81451049-3d54-44c9-9c89-fd336f199313" containerName="pruner" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.302225 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="81451049-3d54-44c9-9c89-fd336f199313" containerName="pruner" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.302469 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="81451049-3d54-44c9-9c89-fd336f199313" containerName="pruner" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.303015 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.310905 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.315642 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.311887 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.453681 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88da617e-cab3-41cc-a2d4-c6e9bd942d55-kube-api-access\") pod \"installer-9-crc\" (UID: \"88da617e-cab3-41cc-a2d4-c6e9bd942d55\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.454030 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/88da617e-cab3-41cc-a2d4-c6e9bd942d55-kubelet-dir\") pod \"installer-9-crc\" (UID: \"88da617e-cab3-41cc-a2d4-c6e9bd942d55\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.454385 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/88da617e-cab3-41cc-a2d4-c6e9bd942d55-var-lock\") pod \"installer-9-crc\" (UID: \"88da617e-cab3-41cc-a2d4-c6e9bd942d55\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.555898 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/88da617e-cab3-41cc-a2d4-c6e9bd942d55-kubelet-dir\") pod \"installer-9-crc\" (UID: \"88da617e-cab3-41cc-a2d4-c6e9bd942d55\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.556524 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/88da617e-cab3-41cc-a2d4-c6e9bd942d55-var-lock\") pod \"installer-9-crc\" (UID: \"88da617e-cab3-41cc-a2d4-c6e9bd942d55\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.556602 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88da617e-cab3-41cc-a2d4-c6e9bd942d55-kube-api-access\") pod \"installer-9-crc\" (UID: \"88da617e-cab3-41cc-a2d4-c6e9bd942d55\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.556074 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/88da617e-cab3-41cc-a2d4-c6e9bd942d55-kubelet-dir\") pod \"installer-9-crc\" (UID: \"88da617e-cab3-41cc-a2d4-c6e9bd942d55\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.556967 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/88da617e-cab3-41cc-a2d4-c6e9bd942d55-var-lock\") pod \"installer-9-crc\" (UID: \"88da617e-cab3-41cc-a2d4-c6e9bd942d55\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.577098 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88da617e-cab3-41cc-a2d4-c6e9bd942d55-kube-api-access\") pod \"installer-9-crc\" (UID: \"88da617e-cab3-41cc-a2d4-c6e9bd942d55\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.632528 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.802194 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 15:37:44 crc kubenswrapper[4739]: W1201 15:37:44.807128 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod88da617e_cab3_41cc_a2d4_c6e9bd942d55.slice/crio-4a8e36b4bf7630fa6f35fb2643bfc929f200662d5ed6cb4f5fd5a5d480563827 WatchSource:0}: Error finding container 4a8e36b4bf7630fa6f35fb2643bfc929f200662d5ed6cb4f5fd5a5d480563827: Status 404 returned error can't find the container with id 4a8e36b4bf7630fa6f35fb2643bfc929f200662d5ed6cb4f5fd5a5d480563827 Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.828473 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"88da617e-cab3-41cc-a2d4-c6e9bd942d55","Type":"ContainerStarted","Data":"4a8e36b4bf7630fa6f35fb2643bfc929f200662d5ed6cb4f5fd5a5d480563827"} Dec 01 15:37:44 crc kubenswrapper[4739]: I1201 15:37:44.831677 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vzjc" event={"ID":"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4","Type":"ContainerStarted","Data":"e0707136eda21e51380324d7ffada4ccb73c59b2454ffa2429d3b625410b1f5b"} Dec 01 15:37:45 crc kubenswrapper[4739]: I1201 15:37:45.151238 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:37:45 crc kubenswrapper[4739]: I1201 15:37:45.151283 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:37:45 crc kubenswrapper[4739]: I1201 15:37:45.232456 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:37:45 crc kubenswrapper[4739]: I1201 15:37:45.556990 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:45 crc kubenswrapper[4739]: I1201 15:37:45.557038 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:45 crc kubenswrapper[4739]: I1201 15:37:45.603949 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:45 crc kubenswrapper[4739]: I1201 15:37:45.837078 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"88da617e-cab3-41cc-a2d4-c6e9bd942d55","Type":"ContainerStarted","Data":"235154dd4175e7712c850a3b181366a50ca5d588300954316c12b1c9e31109cf"} Dec 01 15:37:45 crc kubenswrapper[4739]: I1201 15:37:45.839099 4739 generic.go:334] "Generic (PLEG): container finished" podID="cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" containerID="e0707136eda21e51380324d7ffada4ccb73c59b2454ffa2429d3b625410b1f5b" exitCode=0 Dec 01 15:37:45 crc kubenswrapper[4739]: I1201 15:37:45.839255 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vzjc" event={"ID":"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4","Type":"ContainerDied","Data":"e0707136eda21e51380324d7ffada4ccb73c59b2454ffa2429d3b625410b1f5b"} Dec 01 15:37:45 crc kubenswrapper[4739]: I1201 15:37:45.859296 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.859276589 podStartE2EDuration="1.859276589s" podCreationTimestamp="2025-12-01 15:37:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:37:45.856969148 +0000 UTC m=+167.682715242" watchObservedRunningTime="2025-12-01 15:37:45.859276589 +0000 UTC m=+167.685022683" Dec 01 15:37:45 crc kubenswrapper[4739]: I1201 15:37:45.903747 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:37:45 crc kubenswrapper[4739]: I1201 15:37:45.935269 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:48 crc kubenswrapper[4739]: I1201 15:37:48.460202 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xwdtf"] Dec 01 15:37:48 crc kubenswrapper[4739]: I1201 15:37:48.460703 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xwdtf" podUID="bfbd2a39-0be3-48df-8797-55b4e72603d7" containerName="registry-server" containerID="cri-o://79eee8a23eb277a3572615023fd650c2e620e91a63e18bfeea078a2d2b855155" gracePeriod=2 Dec 01 15:37:49 crc kubenswrapper[4739]: I1201 15:37:49.863992 4739 generic.go:334] "Generic (PLEG): container finished" podID="bfbd2a39-0be3-48df-8797-55b4e72603d7" containerID="79eee8a23eb277a3572615023fd650c2e620e91a63e18bfeea078a2d2b855155" exitCode=0 Dec 01 15:37:49 crc kubenswrapper[4739]: I1201 15:37:49.864041 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwdtf" event={"ID":"bfbd2a39-0be3-48df-8797-55b4e72603d7","Type":"ContainerDied","Data":"79eee8a23eb277a3572615023fd650c2e620e91a63e18bfeea078a2d2b855155"} Dec 01 15:37:49 crc kubenswrapper[4739]: I1201 15:37:49.864384 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwdtf" event={"ID":"bfbd2a39-0be3-48df-8797-55b4e72603d7","Type":"ContainerDied","Data":"6e40d5603e47bdb95b304594728672ddc5f2c4d8ea640a86126921a67f984b2d"} Dec 01 15:37:49 crc kubenswrapper[4739]: I1201 15:37:49.864406 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e40d5603e47bdb95b304594728672ddc5f2c4d8ea640a86126921a67f984b2d" Dec 01 15:37:49 crc kubenswrapper[4739]: I1201 15:37:49.890126 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:50 crc kubenswrapper[4739]: I1201 15:37:50.040249 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfbd2a39-0be3-48df-8797-55b4e72603d7-catalog-content\") pod \"bfbd2a39-0be3-48df-8797-55b4e72603d7\" (UID: \"bfbd2a39-0be3-48df-8797-55b4e72603d7\") " Dec 01 15:37:50 crc kubenswrapper[4739]: I1201 15:37:50.040315 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbz2g\" (UniqueName: \"kubernetes.io/projected/bfbd2a39-0be3-48df-8797-55b4e72603d7-kube-api-access-wbz2g\") pod \"bfbd2a39-0be3-48df-8797-55b4e72603d7\" (UID: \"bfbd2a39-0be3-48df-8797-55b4e72603d7\") " Dec 01 15:37:50 crc kubenswrapper[4739]: I1201 15:37:50.040505 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfbd2a39-0be3-48df-8797-55b4e72603d7-utilities\") pod \"bfbd2a39-0be3-48df-8797-55b4e72603d7\" (UID: \"bfbd2a39-0be3-48df-8797-55b4e72603d7\") " Dec 01 15:37:50 crc kubenswrapper[4739]: I1201 15:37:50.041356 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfbd2a39-0be3-48df-8797-55b4e72603d7-utilities" (OuterVolumeSpecName: "utilities") pod "bfbd2a39-0be3-48df-8797-55b4e72603d7" (UID: "bfbd2a39-0be3-48df-8797-55b4e72603d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:37:50 crc kubenswrapper[4739]: I1201 15:37:50.054241 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfbd2a39-0be3-48df-8797-55b4e72603d7-kube-api-access-wbz2g" (OuterVolumeSpecName: "kube-api-access-wbz2g") pod "bfbd2a39-0be3-48df-8797-55b4e72603d7" (UID: "bfbd2a39-0be3-48df-8797-55b4e72603d7"). InnerVolumeSpecName "kube-api-access-wbz2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:37:50 crc kubenswrapper[4739]: I1201 15:37:50.142274 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfbd2a39-0be3-48df-8797-55b4e72603d7-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:37:50 crc kubenswrapper[4739]: I1201 15:37:50.142310 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbz2g\" (UniqueName: \"kubernetes.io/projected/bfbd2a39-0be3-48df-8797-55b4e72603d7-kube-api-access-wbz2g\") on node \"crc\" DevicePath \"\"" Dec 01 15:37:50 crc kubenswrapper[4739]: I1201 15:37:50.497870 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfbd2a39-0be3-48df-8797-55b4e72603d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bfbd2a39-0be3-48df-8797-55b4e72603d7" (UID: "bfbd2a39-0be3-48df-8797-55b4e72603d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:37:50 crc kubenswrapper[4739]: I1201 15:37:50.547170 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfbd2a39-0be3-48df-8797-55b4e72603d7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:37:50 crc kubenswrapper[4739]: I1201 15:37:50.872000 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwdtf" Dec 01 15:37:50 crc kubenswrapper[4739]: I1201 15:37:50.902324 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xwdtf"] Dec 01 15:37:50 crc kubenswrapper[4739]: I1201 15:37:50.913626 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xwdtf"] Dec 01 15:37:52 crc kubenswrapper[4739]: I1201 15:37:52.484030 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfbd2a39-0be3-48df-8797-55b4e72603d7" path="/var/lib/kubelet/pods/bfbd2a39-0be3-48df-8797-55b4e72603d7/volumes" Dec 01 15:37:54 crc kubenswrapper[4739]: I1201 15:37:54.908029 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5d6l" event={"ID":"44224079-e66e-4e6f-a74e-5595d3a378f2","Type":"ContainerStarted","Data":"f991fde905f3ba278301e739122473fa804290580e821bdae59344c3e570444c"} Dec 01 15:37:54 crc kubenswrapper[4739]: I1201 15:37:54.909904 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9bd8" event={"ID":"16013e7f-96c5-42ab-ad19-f33b1c0e672c","Type":"ContainerStarted","Data":"2275afdc4770f42ebebfa2a2b5a697d4474cb618e851e415b74273e61b6d0dfc"} Dec 01 15:37:54 crc kubenswrapper[4739]: I1201 15:37:54.914671 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2t8q" event={"ID":"68008673-38d1-477e-8963-1f4d069ba0c7","Type":"ContainerStarted","Data":"633b286f9e57248d6b5b5cdd91f6ca1e0af83e0b0a6a595c9a638b91290af0f0"} Dec 01 15:37:54 crc kubenswrapper[4739]: I1201 15:37:54.919983 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwqtt" event={"ID":"5b9930dd-3336-438d-84fc-3a62b6a86e75","Type":"ContainerStarted","Data":"fbd388a660b2093b55c0b4ea8cd1159b870c00492dcee35bc04311bc30aa5f62"} Dec 01 15:37:54 crc kubenswrapper[4739]: I1201 15:37:54.923606 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2bhxd" event={"ID":"4c29aacd-8294-4a12-b340-253e25ebe298","Type":"ContainerStarted","Data":"969043142b34d9921fa068a91692310f3db2784a60f057a0f8dd5dccb19137a7"} Dec 01 15:37:54 crc kubenswrapper[4739]: I1201 15:37:54.925866 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vzjc" event={"ID":"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4","Type":"ContainerStarted","Data":"69ad0151a96195be86ef2057a7e224dabf750cdfcfaf752078ec3f058613a84b"} Dec 01 15:37:55 crc kubenswrapper[4739]: I1201 15:37:55.052458 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8vzjc" podStartSLOduration=2.940477762 podStartE2EDuration="54.052433014s" podCreationTimestamp="2025-12-01 15:37:01 +0000 UTC" firstStartedPulling="2025-12-01 15:37:03.467981039 +0000 UTC m=+125.293727133" lastFinishedPulling="2025-12-01 15:37:54.579936291 +0000 UTC m=+176.405682385" observedRunningTime="2025-12-01 15:37:55.049781552 +0000 UTC m=+176.875527646" watchObservedRunningTime="2025-12-01 15:37:55.052433014 +0000 UTC m=+176.878179128" Dec 01 15:37:55 crc kubenswrapper[4739]: I1201 15:37:55.936289 4739 generic.go:334] "Generic (PLEG): container finished" podID="44224079-e66e-4e6f-a74e-5595d3a378f2" containerID="f991fde905f3ba278301e739122473fa804290580e821bdae59344c3e570444c" exitCode=0 Dec 01 15:37:55 crc kubenswrapper[4739]: I1201 15:37:55.936354 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5d6l" event={"ID":"44224079-e66e-4e6f-a74e-5595d3a378f2","Type":"ContainerDied","Data":"f991fde905f3ba278301e739122473fa804290580e821bdae59344c3e570444c"} Dec 01 15:37:55 crc kubenswrapper[4739]: I1201 15:37:55.942588 4739 generic.go:334] "Generic (PLEG): container finished" podID="16013e7f-96c5-42ab-ad19-f33b1c0e672c" containerID="2275afdc4770f42ebebfa2a2b5a697d4474cb618e851e415b74273e61b6d0dfc" exitCode=0 Dec 01 15:37:55 crc kubenswrapper[4739]: I1201 15:37:55.942634 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9bd8" event={"ID":"16013e7f-96c5-42ab-ad19-f33b1c0e672c","Type":"ContainerDied","Data":"2275afdc4770f42ebebfa2a2b5a697d4474cb618e851e415b74273e61b6d0dfc"} Dec 01 15:37:55 crc kubenswrapper[4739]: I1201 15:37:55.950720 4739 generic.go:334] "Generic (PLEG): container finished" podID="5b9930dd-3336-438d-84fc-3a62b6a86e75" containerID="fbd388a660b2093b55c0b4ea8cd1159b870c00492dcee35bc04311bc30aa5f62" exitCode=0 Dec 01 15:37:55 crc kubenswrapper[4739]: I1201 15:37:55.950824 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwqtt" event={"ID":"5b9930dd-3336-438d-84fc-3a62b6a86e75","Type":"ContainerDied","Data":"fbd388a660b2093b55c0b4ea8cd1159b870c00492dcee35bc04311bc30aa5f62"} Dec 01 15:37:55 crc kubenswrapper[4739]: I1201 15:37:55.967942 4739 generic.go:334] "Generic (PLEG): container finished" podID="68008673-38d1-477e-8963-1f4d069ba0c7" containerID="633b286f9e57248d6b5b5cdd91f6ca1e0af83e0b0a6a595c9a638b91290af0f0" exitCode=0 Dec 01 15:37:55 crc kubenswrapper[4739]: I1201 15:37:55.968267 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2t8q" event={"ID":"68008673-38d1-477e-8963-1f4d069ba0c7","Type":"ContainerDied","Data":"633b286f9e57248d6b5b5cdd91f6ca1e0af83e0b0a6a595c9a638b91290af0f0"} Dec 01 15:37:55 crc kubenswrapper[4739]: I1201 15:37:55.971356 4739 generic.go:334] "Generic (PLEG): container finished" podID="4c29aacd-8294-4a12-b340-253e25ebe298" containerID="969043142b34d9921fa068a91692310f3db2784a60f057a0f8dd5dccb19137a7" exitCode=0 Dec 01 15:37:55 crc kubenswrapper[4739]: I1201 15:37:55.971452 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2bhxd" event={"ID":"4c29aacd-8294-4a12-b340-253e25ebe298","Type":"ContainerDied","Data":"969043142b34d9921fa068a91692310f3db2784a60f057a0f8dd5dccb19137a7"} Dec 01 15:37:56 crc kubenswrapper[4739]: I1201 15:37:56.987798 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5d6l" event={"ID":"44224079-e66e-4e6f-a74e-5595d3a378f2","Type":"ContainerStarted","Data":"617ee75a624932ed3767f42d74925f8a0e7f65dd4132310fbdb8d4fdb6607b8f"} Dec 01 15:37:56 crc kubenswrapper[4739]: I1201 15:37:56.992677 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9bd8" event={"ID":"16013e7f-96c5-42ab-ad19-f33b1c0e672c","Type":"ContainerStarted","Data":"fcacd25f1afb78fae8baf3299dd750617e5e6e0c1d1533477b2c7f47c1928b49"} Dec 01 15:37:57 crc kubenswrapper[4739]: I1201 15:37:57.008636 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z5d6l" podStartSLOduration=2.10737671 podStartE2EDuration="54.008614663s" podCreationTimestamp="2025-12-01 15:37:03 +0000 UTC" firstStartedPulling="2025-12-01 15:37:04.435543848 +0000 UTC m=+126.261289942" lastFinishedPulling="2025-12-01 15:37:56.336781791 +0000 UTC m=+178.162527895" observedRunningTime="2025-12-01 15:37:57.004442493 +0000 UTC m=+178.830188587" watchObservedRunningTime="2025-12-01 15:37:57.008614663 +0000 UTC m=+178.834360777" Dec 01 15:37:57 crc kubenswrapper[4739]: I1201 15:37:57.025845 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l9bd8" podStartSLOduration=3.13143165 podStartE2EDuration="54.025829195s" podCreationTimestamp="2025-12-01 15:37:03 +0000 UTC" firstStartedPulling="2025-12-01 15:37:05.484740837 +0000 UTC m=+127.310486931" lastFinishedPulling="2025-12-01 15:37:56.379138342 +0000 UTC m=+178.204884476" observedRunningTime="2025-12-01 15:37:57.023611386 +0000 UTC m=+178.849357480" watchObservedRunningTime="2025-12-01 15:37:57.025829195 +0000 UTC m=+178.851575289" Dec 01 15:37:58 crc kubenswrapper[4739]: I1201 15:37:58.000924 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2t8q" event={"ID":"68008673-38d1-477e-8963-1f4d069ba0c7","Type":"ContainerStarted","Data":"bf84dbf94e7dfd645a89273ece0f5c49638f053219998dafeff9d33934326e08"} Dec 01 15:37:58 crc kubenswrapper[4739]: I1201 15:37:58.003064 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwqtt" event={"ID":"5b9930dd-3336-438d-84fc-3a62b6a86e75","Type":"ContainerStarted","Data":"72f42d0ff09f68b5d8669e5dc48d5fef6bae5276fd9dc82aee41e25ce11854e9"} Dec 01 15:37:58 crc kubenswrapper[4739]: I1201 15:37:58.005210 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2bhxd" event={"ID":"4c29aacd-8294-4a12-b340-253e25ebe298","Type":"ContainerStarted","Data":"f81b59af7290480dc725ff9319fc892342515e61341f7070b00cfe25c3c2c498"} Dec 01 15:37:58 crc kubenswrapper[4739]: I1201 15:37:58.025332 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r2t8q" podStartSLOduration=3.632237247 podStartE2EDuration="56.025315797s" podCreationTimestamp="2025-12-01 15:37:02 +0000 UTC" firstStartedPulling="2025-12-01 15:37:04.470990729 +0000 UTC m=+126.296736823" lastFinishedPulling="2025-12-01 15:37:56.864069239 +0000 UTC m=+178.689815373" observedRunningTime="2025-12-01 15:37:58.021973143 +0000 UTC m=+179.847719247" watchObservedRunningTime="2025-12-01 15:37:58.025315797 +0000 UTC m=+179.851061891" Dec 01 15:37:58 crc kubenswrapper[4739]: I1201 15:37:58.042089 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nwqtt" podStartSLOduration=3.571680808 podStartE2EDuration="57.042073126s" podCreationTimestamp="2025-12-01 15:37:01 +0000 UTC" firstStartedPulling="2025-12-01 15:37:03.467440714 +0000 UTC m=+125.293186808" lastFinishedPulling="2025-12-01 15:37:56.937833032 +0000 UTC m=+178.763579126" observedRunningTime="2025-12-01 15:37:58.038615498 +0000 UTC m=+179.864361592" watchObservedRunningTime="2025-12-01 15:37:58.042073126 +0000 UTC m=+179.867819220" Dec 01 15:37:58 crc kubenswrapper[4739]: I1201 15:37:58.055325 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2bhxd" podStartSLOduration=4.502752667 podStartE2EDuration="57.055309835s" podCreationTimestamp="2025-12-01 15:37:01 +0000 UTC" firstStartedPulling="2025-12-01 15:37:04.426224321 +0000 UTC m=+126.251970415" lastFinishedPulling="2025-12-01 15:37:56.978781469 +0000 UTC m=+178.804527583" observedRunningTime="2025-12-01 15:37:58.05385862 +0000 UTC m=+179.879604704" watchObservedRunningTime="2025-12-01 15:37:58.055309835 +0000 UTC m=+179.881055929" Dec 01 15:38:01 crc kubenswrapper[4739]: I1201 15:38:01.969478 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:38:01 crc kubenswrapper[4739]: I1201 15:38:01.970662 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:38:02 crc kubenswrapper[4739]: I1201 15:38:02.055160 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:38:02 crc kubenswrapper[4739]: I1201 15:38:02.673555 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:38:02 crc kubenswrapper[4739]: I1201 15:38:02.673627 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:38:02 crc kubenswrapper[4739]: I1201 15:38:02.720583 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:38:02 crc kubenswrapper[4739]: I1201 15:38:02.856562 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:38:02 crc kubenswrapper[4739]: I1201 15:38:02.856627 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:38:02 crc kubenswrapper[4739]: I1201 15:38:02.904183 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:38:03 crc kubenswrapper[4739]: I1201 15:38:03.053531 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:38:03 crc kubenswrapper[4739]: I1201 15:38:03.053608 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:38:03 crc kubenswrapper[4739]: I1201 15:38:03.095015 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:38:03 crc kubenswrapper[4739]: I1201 15:38:03.097946 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:38:03 crc kubenswrapper[4739]: I1201 15:38:03.109327 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:38:03 crc kubenswrapper[4739]: I1201 15:38:03.110319 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:38:03 crc kubenswrapper[4739]: I1201 15:38:03.786624 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:38:03 crc kubenswrapper[4739]: I1201 15:38:03.786678 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:38:03 crc kubenswrapper[4739]: I1201 15:38:03.838478 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:38:04 crc kubenswrapper[4739]: I1201 15:38:04.100807 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:38:04 crc kubenswrapper[4739]: I1201 15:38:04.100911 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:38:04 crc kubenswrapper[4739]: I1201 15:38:04.169983 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:38:04 crc kubenswrapper[4739]: I1201 15:38:04.170044 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:38:04 crc kubenswrapper[4739]: I1201 15:38:04.229596 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:38:05 crc kubenswrapper[4739]: I1201 15:38:05.104410 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:38:05 crc kubenswrapper[4739]: I1201 15:38:05.105984 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 15:38:05 crc kubenswrapper[4739]: I1201 15:38:05.262792 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2bhxd"] Dec 01 15:38:05 crc kubenswrapper[4739]: I1201 15:38:05.468109 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r2t8q"] Dec 01 15:38:05 crc kubenswrapper[4739]: I1201 15:38:05.468376 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r2t8q" podUID="68008673-38d1-477e-8963-1f4d069ba0c7" containerName="registry-server" containerID="cri-o://bf84dbf94e7dfd645a89273ece0f5c49638f053219998dafeff9d33934326e08" gracePeriod=2 Dec 01 15:38:05 crc kubenswrapper[4739]: I1201 15:38:05.518394 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mjqqv"] Dec 01 15:38:06 crc kubenswrapper[4739]: I1201 15:38:06.054280 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2bhxd" podUID="4c29aacd-8294-4a12-b340-253e25ebe298" containerName="registry-server" containerID="cri-o://f81b59af7290480dc725ff9319fc892342515e61341f7070b00cfe25c3c2c498" gracePeriod=2 Dec 01 15:38:07 crc kubenswrapper[4739]: I1201 15:38:07.064288 4739 generic.go:334] "Generic (PLEG): container finished" podID="68008673-38d1-477e-8963-1f4d069ba0c7" containerID="bf84dbf94e7dfd645a89273ece0f5c49638f053219998dafeff9d33934326e08" exitCode=0 Dec 01 15:38:07 crc kubenswrapper[4739]: I1201 15:38:07.064340 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2t8q" event={"ID":"68008673-38d1-477e-8963-1f4d069ba0c7","Type":"ContainerDied","Data":"bf84dbf94e7dfd645a89273ece0f5c49638f053219998dafeff9d33934326e08"} Dec 01 15:38:07 crc kubenswrapper[4739]: I1201 15:38:07.662323 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l9bd8"] Dec 01 15:38:07 crc kubenswrapper[4739]: I1201 15:38:07.663055 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l9bd8" podUID="16013e7f-96c5-42ab-ad19-f33b1c0e672c" containerName="registry-server" containerID="cri-o://fcacd25f1afb78fae8baf3299dd750617e5e6e0c1d1533477b2c7f47c1928b49" gracePeriod=2 Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.007318 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.012031 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.074827 4739 generic.go:334] "Generic (PLEG): container finished" podID="16013e7f-96c5-42ab-ad19-f33b1c0e672c" containerID="fcacd25f1afb78fae8baf3299dd750617e5e6e0c1d1533477b2c7f47c1928b49" exitCode=0 Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.074947 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9bd8" event={"ID":"16013e7f-96c5-42ab-ad19-f33b1c0e672c","Type":"ContainerDied","Data":"fcacd25f1afb78fae8baf3299dd750617e5e6e0c1d1533477b2c7f47c1928b49"} Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.078890 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2t8q" event={"ID":"68008673-38d1-477e-8963-1f4d069ba0c7","Type":"ContainerDied","Data":"cbf676fdf19bb60a7c099dcbd720db64727db65ab0a771bf27023922ad5aa588"} Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.078963 4739 scope.go:117] "RemoveContainer" containerID="bf84dbf94e7dfd645a89273ece0f5c49638f053219998dafeff9d33934326e08" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.079320 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2t8q" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.085091 4739 generic.go:334] "Generic (PLEG): container finished" podID="4c29aacd-8294-4a12-b340-253e25ebe298" containerID="f81b59af7290480dc725ff9319fc892342515e61341f7070b00cfe25c3c2c498" exitCode=0 Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.085135 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2bhxd" event={"ID":"4c29aacd-8294-4a12-b340-253e25ebe298","Type":"ContainerDied","Data":"f81b59af7290480dc725ff9319fc892342515e61341f7070b00cfe25c3c2c498"} Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.085171 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2bhxd" event={"ID":"4c29aacd-8294-4a12-b340-253e25ebe298","Type":"ContainerDied","Data":"9c41fcbec5dbcaf4c2a7bf7ed78d39fe6b14cf6cbfcbc684a6a7cc61b8a91815"} Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.085245 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2bhxd" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.097987 4739 scope.go:117] "RemoveContainer" containerID="633b286f9e57248d6b5b5cdd91f6ca1e0af83e0b0a6a595c9a638b91290af0f0" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.098684 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpjkw\" (UniqueName: \"kubernetes.io/projected/68008673-38d1-477e-8963-1f4d069ba0c7-kube-api-access-dpjkw\") pod \"68008673-38d1-477e-8963-1f4d069ba0c7\" (UID: \"68008673-38d1-477e-8963-1f4d069ba0c7\") " Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.099238 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5dkv\" (UniqueName: \"kubernetes.io/projected/4c29aacd-8294-4a12-b340-253e25ebe298-kube-api-access-k5dkv\") pod \"4c29aacd-8294-4a12-b340-253e25ebe298\" (UID: \"4c29aacd-8294-4a12-b340-253e25ebe298\") " Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.099331 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c29aacd-8294-4a12-b340-253e25ebe298-utilities\") pod \"4c29aacd-8294-4a12-b340-253e25ebe298\" (UID: \"4c29aacd-8294-4a12-b340-253e25ebe298\") " Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.099370 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68008673-38d1-477e-8963-1f4d069ba0c7-catalog-content\") pod \"68008673-38d1-477e-8963-1f4d069ba0c7\" (UID: \"68008673-38d1-477e-8963-1f4d069ba0c7\") " Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.099405 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68008673-38d1-477e-8963-1f4d069ba0c7-utilities\") pod \"68008673-38d1-477e-8963-1f4d069ba0c7\" (UID: \"68008673-38d1-477e-8963-1f4d069ba0c7\") " Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.099461 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c29aacd-8294-4a12-b340-253e25ebe298-catalog-content\") pod \"4c29aacd-8294-4a12-b340-253e25ebe298\" (UID: \"4c29aacd-8294-4a12-b340-253e25ebe298\") " Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.100476 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c29aacd-8294-4a12-b340-253e25ebe298-utilities" (OuterVolumeSpecName: "utilities") pod "4c29aacd-8294-4a12-b340-253e25ebe298" (UID: "4c29aacd-8294-4a12-b340-253e25ebe298"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.100571 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68008673-38d1-477e-8963-1f4d069ba0c7-utilities" (OuterVolumeSpecName: "utilities") pod "68008673-38d1-477e-8963-1f4d069ba0c7" (UID: "68008673-38d1-477e-8963-1f4d069ba0c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.108333 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c29aacd-8294-4a12-b340-253e25ebe298-kube-api-access-k5dkv" (OuterVolumeSpecName: "kube-api-access-k5dkv") pod "4c29aacd-8294-4a12-b340-253e25ebe298" (UID: "4c29aacd-8294-4a12-b340-253e25ebe298"). InnerVolumeSpecName "kube-api-access-k5dkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.113239 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68008673-38d1-477e-8963-1f4d069ba0c7-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.113267 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5dkv\" (UniqueName: \"kubernetes.io/projected/4c29aacd-8294-4a12-b340-253e25ebe298-kube-api-access-k5dkv\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.113281 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c29aacd-8294-4a12-b340-253e25ebe298-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.113722 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68008673-38d1-477e-8963-1f4d069ba0c7-kube-api-access-dpjkw" (OuterVolumeSpecName: "kube-api-access-dpjkw") pod "68008673-38d1-477e-8963-1f4d069ba0c7" (UID: "68008673-38d1-477e-8963-1f4d069ba0c7"). InnerVolumeSpecName "kube-api-access-dpjkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.115992 4739 scope.go:117] "RemoveContainer" containerID="1e5a101dc84bb98fde41bc27c306513659c4dd027bba302f99733d9728a80dd3" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.148261 4739 scope.go:117] "RemoveContainer" containerID="f81b59af7290480dc725ff9319fc892342515e61341f7070b00cfe25c3c2c498" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.163714 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68008673-38d1-477e-8963-1f4d069ba0c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68008673-38d1-477e-8963-1f4d069ba0c7" (UID: "68008673-38d1-477e-8963-1f4d069ba0c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.168956 4739 scope.go:117] "RemoveContainer" containerID="969043142b34d9921fa068a91692310f3db2784a60f057a0f8dd5dccb19137a7" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.173860 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c29aacd-8294-4a12-b340-253e25ebe298-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c29aacd-8294-4a12-b340-253e25ebe298" (UID: "4c29aacd-8294-4a12-b340-253e25ebe298"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.187840 4739 scope.go:117] "RemoveContainer" containerID="13732b0fc46b3e3efce48e313887e8adfb880a9a365daff07a6152e09abcafa9" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.214980 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68008673-38d1-477e-8963-1f4d069ba0c7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.215013 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c29aacd-8294-4a12-b340-253e25ebe298-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.215023 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpjkw\" (UniqueName: \"kubernetes.io/projected/68008673-38d1-477e-8963-1f4d069ba0c7-kube-api-access-dpjkw\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.216874 4739 scope.go:117] "RemoveContainer" containerID="f81b59af7290480dc725ff9319fc892342515e61341f7070b00cfe25c3c2c498" Dec 01 15:38:08 crc kubenswrapper[4739]: E1201 15:38:08.217255 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f81b59af7290480dc725ff9319fc892342515e61341f7070b00cfe25c3c2c498\": container with ID starting with f81b59af7290480dc725ff9319fc892342515e61341f7070b00cfe25c3c2c498 not found: ID does not exist" containerID="f81b59af7290480dc725ff9319fc892342515e61341f7070b00cfe25c3c2c498" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.217283 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f81b59af7290480dc725ff9319fc892342515e61341f7070b00cfe25c3c2c498"} err="failed to get container status \"f81b59af7290480dc725ff9319fc892342515e61341f7070b00cfe25c3c2c498\": rpc error: code = NotFound desc = could not find container \"f81b59af7290480dc725ff9319fc892342515e61341f7070b00cfe25c3c2c498\": container with ID starting with f81b59af7290480dc725ff9319fc892342515e61341f7070b00cfe25c3c2c498 not found: ID does not exist" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.217319 4739 scope.go:117] "RemoveContainer" containerID="969043142b34d9921fa068a91692310f3db2784a60f057a0f8dd5dccb19137a7" Dec 01 15:38:08 crc kubenswrapper[4739]: E1201 15:38:08.217611 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"969043142b34d9921fa068a91692310f3db2784a60f057a0f8dd5dccb19137a7\": container with ID starting with 969043142b34d9921fa068a91692310f3db2784a60f057a0f8dd5dccb19137a7 not found: ID does not exist" containerID="969043142b34d9921fa068a91692310f3db2784a60f057a0f8dd5dccb19137a7" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.217631 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"969043142b34d9921fa068a91692310f3db2784a60f057a0f8dd5dccb19137a7"} err="failed to get container status \"969043142b34d9921fa068a91692310f3db2784a60f057a0f8dd5dccb19137a7\": rpc error: code = NotFound desc = could not find container \"969043142b34d9921fa068a91692310f3db2784a60f057a0f8dd5dccb19137a7\": container with ID starting with 969043142b34d9921fa068a91692310f3db2784a60f057a0f8dd5dccb19137a7 not found: ID does not exist" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.217645 4739 scope.go:117] "RemoveContainer" containerID="13732b0fc46b3e3efce48e313887e8adfb880a9a365daff07a6152e09abcafa9" Dec 01 15:38:08 crc kubenswrapper[4739]: E1201 15:38:08.217950 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13732b0fc46b3e3efce48e313887e8adfb880a9a365daff07a6152e09abcafa9\": container with ID starting with 13732b0fc46b3e3efce48e313887e8adfb880a9a365daff07a6152e09abcafa9 not found: ID does not exist" containerID="13732b0fc46b3e3efce48e313887e8adfb880a9a365daff07a6152e09abcafa9" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.217968 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13732b0fc46b3e3efce48e313887e8adfb880a9a365daff07a6152e09abcafa9"} err="failed to get container status \"13732b0fc46b3e3efce48e313887e8adfb880a9a365daff07a6152e09abcafa9\": rpc error: code = NotFound desc = could not find container \"13732b0fc46b3e3efce48e313887e8adfb880a9a365daff07a6152e09abcafa9\": container with ID starting with 13732b0fc46b3e3efce48e313887e8adfb880a9a365daff07a6152e09abcafa9 not found: ID does not exist" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.414114 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2bhxd"] Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.418700 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2bhxd"] Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.426101 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r2t8q"] Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.428492 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r2t8q"] Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.486746 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c29aacd-8294-4a12-b340-253e25ebe298" path="/var/lib/kubelet/pods/4c29aacd-8294-4a12-b340-253e25ebe298/volumes" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.488594 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68008673-38d1-477e-8963-1f4d069ba0c7" path="/var/lib/kubelet/pods/68008673-38d1-477e-8963-1f4d069ba0c7/volumes" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.778871 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.922942 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grgx6\" (UniqueName: \"kubernetes.io/projected/16013e7f-96c5-42ab-ad19-f33b1c0e672c-kube-api-access-grgx6\") pod \"16013e7f-96c5-42ab-ad19-f33b1c0e672c\" (UID: \"16013e7f-96c5-42ab-ad19-f33b1c0e672c\") " Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.923017 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16013e7f-96c5-42ab-ad19-f33b1c0e672c-catalog-content\") pod \"16013e7f-96c5-42ab-ad19-f33b1c0e672c\" (UID: \"16013e7f-96c5-42ab-ad19-f33b1c0e672c\") " Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.923063 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16013e7f-96c5-42ab-ad19-f33b1c0e672c-utilities\") pod \"16013e7f-96c5-42ab-ad19-f33b1c0e672c\" (UID: \"16013e7f-96c5-42ab-ad19-f33b1c0e672c\") " Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.924252 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16013e7f-96c5-42ab-ad19-f33b1c0e672c-utilities" (OuterVolumeSpecName: "utilities") pod "16013e7f-96c5-42ab-ad19-f33b1c0e672c" (UID: "16013e7f-96c5-42ab-ad19-f33b1c0e672c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.927217 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16013e7f-96c5-42ab-ad19-f33b1c0e672c-kube-api-access-grgx6" (OuterVolumeSpecName: "kube-api-access-grgx6") pod "16013e7f-96c5-42ab-ad19-f33b1c0e672c" (UID: "16013e7f-96c5-42ab-ad19-f33b1c0e672c"). InnerVolumeSpecName "kube-api-access-grgx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:38:08 crc kubenswrapper[4739]: I1201 15:38:08.949169 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16013e7f-96c5-42ab-ad19-f33b1c0e672c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "16013e7f-96c5-42ab-ad19-f33b1c0e672c" (UID: "16013e7f-96c5-42ab-ad19-f33b1c0e672c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:38:09 crc kubenswrapper[4739]: I1201 15:38:09.024604 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grgx6\" (UniqueName: \"kubernetes.io/projected/16013e7f-96c5-42ab-ad19-f33b1c0e672c-kube-api-access-grgx6\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:09 crc kubenswrapper[4739]: I1201 15:38:09.024667 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16013e7f-96c5-42ab-ad19-f33b1c0e672c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:09 crc kubenswrapper[4739]: I1201 15:38:09.024686 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16013e7f-96c5-42ab-ad19-f33b1c0e672c-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:09 crc kubenswrapper[4739]: I1201 15:38:09.095007 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9bd8" event={"ID":"16013e7f-96c5-42ab-ad19-f33b1c0e672c","Type":"ContainerDied","Data":"bff74b14c9462f5aee08189d4b39fbbdf2f898dba75073493cd189239f966cca"} Dec 01 15:38:09 crc kubenswrapper[4739]: I1201 15:38:09.095057 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l9bd8" Dec 01 15:38:09 crc kubenswrapper[4739]: I1201 15:38:09.095064 4739 scope.go:117] "RemoveContainer" containerID="fcacd25f1afb78fae8baf3299dd750617e5e6e0c1d1533477b2c7f47c1928b49" Dec 01 15:38:09 crc kubenswrapper[4739]: I1201 15:38:09.114594 4739 scope.go:117] "RemoveContainer" containerID="2275afdc4770f42ebebfa2a2b5a697d4474cb618e851e415b74273e61b6d0dfc" Dec 01 15:38:09 crc kubenswrapper[4739]: I1201 15:38:09.145600 4739 scope.go:117] "RemoveContainer" containerID="943787895eda2bfcbe4ed1faccdda17dad595de3d8ccb0c68ae8cf215935e2c7" Dec 01 15:38:09 crc kubenswrapper[4739]: I1201 15:38:09.184774 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l9bd8"] Dec 01 15:38:09 crc kubenswrapper[4739]: I1201 15:38:09.188021 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l9bd8"] Dec 01 15:38:09 crc kubenswrapper[4739]: I1201 15:38:09.621620 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:38:09 crc kubenswrapper[4739]: I1201 15:38:09.621680 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:38:10 crc kubenswrapper[4739]: I1201 15:38:10.493577 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16013e7f-96c5-42ab-ad19-f33b1c0e672c" path="/var/lib/kubelet/pods/16013e7f-96c5-42ab-ad19-f33b1c0e672c/volumes" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.859884 4739 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.861166 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfbd2a39-0be3-48df-8797-55b4e72603d7" containerName="registry-server" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861194 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfbd2a39-0be3-48df-8797-55b4e72603d7" containerName="registry-server" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.861231 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c29aacd-8294-4a12-b340-253e25ebe298" containerName="registry-server" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861244 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c29aacd-8294-4a12-b340-253e25ebe298" containerName="registry-server" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.861263 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfbd2a39-0be3-48df-8797-55b4e72603d7" containerName="extract-content" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861275 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfbd2a39-0be3-48df-8797-55b4e72603d7" containerName="extract-content" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.861295 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c29aacd-8294-4a12-b340-253e25ebe298" containerName="extract-content" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861307 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c29aacd-8294-4a12-b340-253e25ebe298" containerName="extract-content" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.861327 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16013e7f-96c5-42ab-ad19-f33b1c0e672c" containerName="extract-content" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861339 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="16013e7f-96c5-42ab-ad19-f33b1c0e672c" containerName="extract-content" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.861359 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16013e7f-96c5-42ab-ad19-f33b1c0e672c" containerName="extract-utilities" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861371 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="16013e7f-96c5-42ab-ad19-f33b1c0e672c" containerName="extract-utilities" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.861390 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfbd2a39-0be3-48df-8797-55b4e72603d7" containerName="extract-utilities" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861403 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfbd2a39-0be3-48df-8797-55b4e72603d7" containerName="extract-utilities" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.861453 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68008673-38d1-477e-8963-1f4d069ba0c7" containerName="extract-content" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861469 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="68008673-38d1-477e-8963-1f4d069ba0c7" containerName="extract-content" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.861484 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16013e7f-96c5-42ab-ad19-f33b1c0e672c" containerName="registry-server" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861496 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="16013e7f-96c5-42ab-ad19-f33b1c0e672c" containerName="registry-server" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.861510 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c29aacd-8294-4a12-b340-253e25ebe298" containerName="extract-utilities" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861522 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c29aacd-8294-4a12-b340-253e25ebe298" containerName="extract-utilities" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.861545 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68008673-38d1-477e-8963-1f4d069ba0c7" containerName="extract-utilities" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861557 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="68008673-38d1-477e-8963-1f4d069ba0c7" containerName="extract-utilities" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.861581 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68008673-38d1-477e-8963-1f4d069ba0c7" containerName="registry-server" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861592 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="68008673-38d1-477e-8963-1f4d069ba0c7" containerName="registry-server" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861772 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="16013e7f-96c5-42ab-ad19-f33b1c0e672c" containerName="registry-server" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861800 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c29aacd-8294-4a12-b340-253e25ebe298" containerName="registry-server" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861817 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfbd2a39-0be3-48df-8797-55b4e72603d7" containerName="registry-server" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.861833 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="68008673-38d1-477e-8963-1f4d069ba0c7" containerName="registry-server" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.862370 4739 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.862620 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.862961 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da" gracePeriod=15 Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.863054 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41" gracePeriod=15 Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.863076 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe" gracePeriod=15 Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.863165 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7" gracePeriod=15 Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.863140 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786" gracePeriod=15 Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.864252 4739 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.864661 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.864711 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.864739 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.864757 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.864781 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.864798 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.864828 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.864844 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.864871 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.864888 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.864913 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.864929 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 15:38:22 crc kubenswrapper[4739]: E1201 15:38:22.864963 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.864979 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.865205 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.865241 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.865271 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.865291 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.865315 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 15:38:22 crc kubenswrapper[4739]: I1201 15:38:22.865344 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.013394 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.013484 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.013533 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.013665 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.013738 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.013792 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.013829 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.013876 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115081 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115126 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115152 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115178 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115196 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115246 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115266 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115311 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115327 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115331 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115350 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115375 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115324 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115446 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115527 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.115593 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.189636 4739 generic.go:334] "Generic (PLEG): container finished" podID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" containerID="235154dd4175e7712c850a3b181366a50ca5d588300954316c12b1c9e31109cf" exitCode=0 Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.189741 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"88da617e-cab3-41cc-a2d4-c6e9bd942d55","Type":"ContainerDied","Data":"235154dd4175e7712c850a3b181366a50ca5d588300954316c12b1c9e31109cf"} Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.190991 4739 status_manager.go:851] "Failed to get status for pod" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.191488 4739 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.192265 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.194147 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.195370 4739 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41" exitCode=0 Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.195406 4739 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe" exitCode=0 Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.195461 4739 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786" exitCode=0 Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.195482 4739 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7" exitCode=2 Dec 01 15:38:23 crc kubenswrapper[4739]: I1201 15:38:23.195515 4739 scope.go:117] "RemoveContainer" containerID="ad2c4dd9e1a0b4b7af88e8dc2a7d3f1e97a13e14863e2d0abc61b2c3a2ca6fe4" Dec 01 15:38:24 crc kubenswrapper[4739]: I1201 15:38:24.205193 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 15:38:24 crc kubenswrapper[4739]: I1201 15:38:24.530364 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 15:38:24 crc kubenswrapper[4739]: I1201 15:38:24.531654 4739 status_manager.go:851] "Failed to get status for pod" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:24 crc kubenswrapper[4739]: I1201 15:38:24.634462 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/88da617e-cab3-41cc-a2d4-c6e9bd942d55-var-lock\") pod \"88da617e-cab3-41cc-a2d4-c6e9bd942d55\" (UID: \"88da617e-cab3-41cc-a2d4-c6e9bd942d55\") " Dec 01 15:38:24 crc kubenswrapper[4739]: I1201 15:38:24.634539 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/88da617e-cab3-41cc-a2d4-c6e9bd942d55-kubelet-dir\") pod \"88da617e-cab3-41cc-a2d4-c6e9bd942d55\" (UID: \"88da617e-cab3-41cc-a2d4-c6e9bd942d55\") " Dec 01 15:38:24 crc kubenswrapper[4739]: I1201 15:38:24.634742 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88da617e-cab3-41cc-a2d4-c6e9bd942d55-kube-api-access\") pod \"88da617e-cab3-41cc-a2d4-c6e9bd942d55\" (UID: \"88da617e-cab3-41cc-a2d4-c6e9bd942d55\") " Dec 01 15:38:24 crc kubenswrapper[4739]: I1201 15:38:24.634682 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88da617e-cab3-41cc-a2d4-c6e9bd942d55-var-lock" (OuterVolumeSpecName: "var-lock") pod "88da617e-cab3-41cc-a2d4-c6e9bd942d55" (UID: "88da617e-cab3-41cc-a2d4-c6e9bd942d55"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:38:24 crc kubenswrapper[4739]: I1201 15:38:24.634847 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88da617e-cab3-41cc-a2d4-c6e9bd942d55-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "88da617e-cab3-41cc-a2d4-c6e9bd942d55" (UID: "88da617e-cab3-41cc-a2d4-c6e9bd942d55"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:38:24 crc kubenswrapper[4739]: I1201 15:38:24.635274 4739 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/88da617e-cab3-41cc-a2d4-c6e9bd942d55-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:24 crc kubenswrapper[4739]: I1201 15:38:24.635305 4739 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/88da617e-cab3-41cc-a2d4-c6e9bd942d55-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:24 crc kubenswrapper[4739]: I1201 15:38:24.640490 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88da617e-cab3-41cc-a2d4-c6e9bd942d55-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "88da617e-cab3-41cc-a2d4-c6e9bd942d55" (UID: "88da617e-cab3-41cc-a2d4-c6e9bd942d55"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:38:24 crc kubenswrapper[4739]: I1201 15:38:24.736633 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88da617e-cab3-41cc-a2d4-c6e9bd942d55-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.215063 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"88da617e-cab3-41cc-a2d4-c6e9bd942d55","Type":"ContainerDied","Data":"4a8e36b4bf7630fa6f35fb2643bfc929f200662d5ed6cb4f5fd5a5d480563827"} Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.215714 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a8e36b4bf7630fa6f35fb2643bfc929f200662d5ed6cb4f5fd5a5d480563827" Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.215160 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.221216 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.222090 4739 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da" exitCode=0 Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.241882 4739 status_manager.go:851] "Failed to get status for pod" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.272203 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.274240 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.274927 4739 status_manager.go:851] "Failed to get status for pod" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.275409 4739 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.446163 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.446314 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.446345 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.446415 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.446454 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.446569 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.446750 4739 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.446776 4739 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:25 crc kubenswrapper[4739]: I1201 15:38:25.446791 4739 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:26 crc kubenswrapper[4739]: E1201 15:38:26.034850 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:38:26Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:38:26Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:38:26Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:38:26Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:26 crc kubenswrapper[4739]: E1201 15:38:26.035465 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:26 crc kubenswrapper[4739]: E1201 15:38:26.036803 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:26 crc kubenswrapper[4739]: E1201 15:38:26.037025 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:26 crc kubenswrapper[4739]: E1201 15:38:26.037216 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:26 crc kubenswrapper[4739]: E1201 15:38:26.037230 4739 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 15:38:26 crc kubenswrapper[4739]: I1201 15:38:26.235916 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 15:38:26 crc kubenswrapper[4739]: I1201 15:38:26.237541 4739 scope.go:117] "RemoveContainer" containerID="dcb96469e1771698a650012471f719800c84c44beb9f0587d14aeae4194b1d41" Dec 01 15:38:26 crc kubenswrapper[4739]: I1201 15:38:26.237656 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:26 crc kubenswrapper[4739]: I1201 15:38:26.256043 4739 scope.go:117] "RemoveContainer" containerID="16f852126f0b7802c7c31379b2975cd450296a0aeefcc8ab01d9d7e85756a2fe" Dec 01 15:38:26 crc kubenswrapper[4739]: I1201 15:38:26.257441 4739 status_manager.go:851] "Failed to get status for pod" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:26 crc kubenswrapper[4739]: I1201 15:38:26.257834 4739 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:26 crc kubenswrapper[4739]: I1201 15:38:26.272321 4739 scope.go:117] "RemoveContainer" containerID="9357d8db225494610088949708c861b46836ee6ccc4f637b148889ee04495786" Dec 01 15:38:26 crc kubenswrapper[4739]: I1201 15:38:26.288770 4739 scope.go:117] "RemoveContainer" containerID="1db259ff1de2aa4cc3d8de250eea63656e895420b75eb13dc72c1fd81273dba7" Dec 01 15:38:26 crc kubenswrapper[4739]: I1201 15:38:26.303045 4739 scope.go:117] "RemoveContainer" containerID="3b9a83f70a3160f4241609196c7bc11b192e8a4c0fd1064fe16e240d0bd415da" Dec 01 15:38:26 crc kubenswrapper[4739]: I1201 15:38:26.319256 4739 scope.go:117] "RemoveContainer" containerID="8df04cbc748cdd93ea85dd0d9c1f50d6cec8a9353f127167a76f3803c8342089" Dec 01 15:38:26 crc kubenswrapper[4739]: I1201 15:38:26.489281 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 01 15:38:27 crc kubenswrapper[4739]: E1201 15:38:27.924135 4739 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.106:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:27 crc kubenswrapper[4739]: I1201 15:38:27.925860 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:27 crc kubenswrapper[4739]: W1201 15:38:27.951934 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-0006d6933b5fd87dfefbdabee8143e730ee7fada364e6f198fe75f0a66e1e7ed WatchSource:0}: Error finding container 0006d6933b5fd87dfefbdabee8143e730ee7fada364e6f198fe75f0a66e1e7ed: Status 404 returned error can't find the container with id 0006d6933b5fd87dfefbdabee8143e730ee7fada364e6f198fe75f0a66e1e7ed Dec 01 15:38:27 crc kubenswrapper[4739]: E1201 15:38:27.955525 4739 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.106:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d218a8d0bedc3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 15:38:27.954675139 +0000 UTC m=+209.780421253,LastTimestamp:2025-12-01 15:38:27.954675139 +0000 UTC m=+209.780421253,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 15:38:28 crc kubenswrapper[4739]: I1201 15:38:28.256583 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"0006d6933b5fd87dfefbdabee8143e730ee7fada364e6f198fe75f0a66e1e7ed"} Dec 01 15:38:28 crc kubenswrapper[4739]: I1201 15:38:28.486308 4739 status_manager.go:851] "Failed to get status for pod" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:29 crc kubenswrapper[4739]: I1201 15:38:29.267849 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"4161b4b7e256ffacaeceaa35d3090c913e8f2798ffd1a67318edfed1d610740f"} Dec 01 15:38:29 crc kubenswrapper[4739]: I1201 15:38:29.268778 4739 status_manager.go:851] "Failed to get status for pod" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:29 crc kubenswrapper[4739]: E1201 15:38:29.268954 4739 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.106:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:30 crc kubenswrapper[4739]: E1201 15:38:30.246919 4739 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:30 crc kubenswrapper[4739]: E1201 15:38:30.247822 4739 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:30 crc kubenswrapper[4739]: E1201 15:38:30.248405 4739 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:30 crc kubenswrapper[4739]: E1201 15:38:30.249080 4739 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:30 crc kubenswrapper[4739]: E1201 15:38:30.249822 4739 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:30 crc kubenswrapper[4739]: I1201 15:38:30.249907 4739 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 01 15:38:30 crc kubenswrapper[4739]: E1201 15:38:30.250463 4739 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" interval="200ms" Dec 01 15:38:30 crc kubenswrapper[4739]: E1201 15:38:30.276284 4739 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.106:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:38:30 crc kubenswrapper[4739]: E1201 15:38:30.452312 4739 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" interval="400ms" Dec 01 15:38:30 crc kubenswrapper[4739]: I1201 15:38:30.558820 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" podUID="840099a9-3ba2-4c14-bcac-de2dcf5c5d95" containerName="oauth-openshift" containerID="cri-o://d9b7ff277dd4af7f73739de635e00cfddfcbb79d18a745ae037985af277022f8" gracePeriod=15 Dec 01 15:38:30 crc kubenswrapper[4739]: E1201 15:38:30.853372 4739 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" interval="800ms" Dec 01 15:38:31 crc kubenswrapper[4739]: E1201 15:38:31.017593 4739 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.106:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d218a8d0bedc3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 15:38:27.954675139 +0000 UTC m=+209.780421253,LastTimestamp:2025-12-01 15:38:27.954675139 +0000 UTC m=+209.780421253,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.038151 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.038731 4739 status_manager.go:851] "Failed to get status for pod" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.039204 4739 status_manager.go:851] "Failed to get status for pod" podUID="840099a9-3ba2-4c14-bcac-de2dcf5c5d95" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-mjqqv\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.231237 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-idp-0-file-data\") pod \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.231587 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-login\") pod \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.231653 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-cliconfig\") pod \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.231688 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-service-ca\") pod \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.231743 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-session\") pod \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.231783 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-audit-dir\") pod \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.231819 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-audit-policies\") pod \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.231882 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-serving-cert\") pod \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.231918 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25rwc\" (UniqueName: \"kubernetes.io/projected/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-kube-api-access-25rwc\") pod \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.231945 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-ocp-branding-template\") pod \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.231982 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-error\") pod \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.232070 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-provider-selection\") pod \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.232154 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-trusted-ca-bundle\") pod \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.232199 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-router-certs\") pod \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\" (UID: \"840099a9-3ba2-4c14-bcac-de2dcf5c5d95\") " Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.232454 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "840099a9-3ba2-4c14-bcac-de2dcf5c5d95" (UID: "840099a9-3ba2-4c14-bcac-de2dcf5c5d95"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.232710 4739 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.232812 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "840099a9-3ba2-4c14-bcac-de2dcf5c5d95" (UID: "840099a9-3ba2-4c14-bcac-de2dcf5c5d95"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.232827 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "840099a9-3ba2-4c14-bcac-de2dcf5c5d95" (UID: "840099a9-3ba2-4c14-bcac-de2dcf5c5d95"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.233558 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "840099a9-3ba2-4c14-bcac-de2dcf5c5d95" (UID: "840099a9-3ba2-4c14-bcac-de2dcf5c5d95"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.233724 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "840099a9-3ba2-4c14-bcac-de2dcf5c5d95" (UID: "840099a9-3ba2-4c14-bcac-de2dcf5c5d95"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.239596 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "840099a9-3ba2-4c14-bcac-de2dcf5c5d95" (UID: "840099a9-3ba2-4c14-bcac-de2dcf5c5d95"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.240041 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "840099a9-3ba2-4c14-bcac-de2dcf5c5d95" (UID: "840099a9-3ba2-4c14-bcac-de2dcf5c5d95"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.240324 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-kube-api-access-25rwc" (OuterVolumeSpecName: "kube-api-access-25rwc") pod "840099a9-3ba2-4c14-bcac-de2dcf5c5d95" (UID: "840099a9-3ba2-4c14-bcac-de2dcf5c5d95"). InnerVolumeSpecName "kube-api-access-25rwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.240407 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "840099a9-3ba2-4c14-bcac-de2dcf5c5d95" (UID: "840099a9-3ba2-4c14-bcac-de2dcf5c5d95"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.241106 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "840099a9-3ba2-4c14-bcac-de2dcf5c5d95" (UID: "840099a9-3ba2-4c14-bcac-de2dcf5c5d95"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.241837 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "840099a9-3ba2-4c14-bcac-de2dcf5c5d95" (UID: "840099a9-3ba2-4c14-bcac-de2dcf5c5d95"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.241914 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "840099a9-3ba2-4c14-bcac-de2dcf5c5d95" (UID: "840099a9-3ba2-4c14-bcac-de2dcf5c5d95"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.243993 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "840099a9-3ba2-4c14-bcac-de2dcf5c5d95" (UID: "840099a9-3ba2-4c14-bcac-de2dcf5c5d95"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.244573 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "840099a9-3ba2-4c14-bcac-de2dcf5c5d95" (UID: "840099a9-3ba2-4c14-bcac-de2dcf5c5d95"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.283604 4739 generic.go:334] "Generic (PLEG): container finished" podID="840099a9-3ba2-4c14-bcac-de2dcf5c5d95" containerID="d9b7ff277dd4af7f73739de635e00cfddfcbb79d18a745ae037985af277022f8" exitCode=0 Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.283665 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" event={"ID":"840099a9-3ba2-4c14-bcac-de2dcf5c5d95","Type":"ContainerDied","Data":"d9b7ff277dd4af7f73739de635e00cfddfcbb79d18a745ae037985af277022f8"} Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.283698 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.283727 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" event={"ID":"840099a9-3ba2-4c14-bcac-de2dcf5c5d95","Type":"ContainerDied","Data":"901a5489cc6fb1ae716c95d51a6ed8f16b05fc791666b45b6f1f741b2700b48a"} Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.283757 4739 scope.go:117] "RemoveContainer" containerID="d9b7ff277dd4af7f73739de635e00cfddfcbb79d18a745ae037985af277022f8" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.284696 4739 status_manager.go:851] "Failed to get status for pod" podUID="840099a9-3ba2-4c14-bcac-de2dcf5c5d95" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-mjqqv\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.285365 4739 status_manager.go:851] "Failed to get status for pod" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.311781 4739 scope.go:117] "RemoveContainer" containerID="d9b7ff277dd4af7f73739de635e00cfddfcbb79d18a745ae037985af277022f8" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.312183 4739 status_manager.go:851] "Failed to get status for pod" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:31 crc kubenswrapper[4739]: E1201 15:38:31.312221 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9b7ff277dd4af7f73739de635e00cfddfcbb79d18a745ae037985af277022f8\": container with ID starting with d9b7ff277dd4af7f73739de635e00cfddfcbb79d18a745ae037985af277022f8 not found: ID does not exist" containerID="d9b7ff277dd4af7f73739de635e00cfddfcbb79d18a745ae037985af277022f8" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.312260 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9b7ff277dd4af7f73739de635e00cfddfcbb79d18a745ae037985af277022f8"} err="failed to get container status \"d9b7ff277dd4af7f73739de635e00cfddfcbb79d18a745ae037985af277022f8\": rpc error: code = NotFound desc = could not find container \"d9b7ff277dd4af7f73739de635e00cfddfcbb79d18a745ae037985af277022f8\": container with ID starting with d9b7ff277dd4af7f73739de635e00cfddfcbb79d18a745ae037985af277022f8 not found: ID does not exist" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.312776 4739 status_manager.go:851] "Failed to get status for pod" podUID="840099a9-3ba2-4c14-bcac-de2dcf5c5d95" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-mjqqv\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.333221 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.333253 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25rwc\" (UniqueName: \"kubernetes.io/projected/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-kube-api-access-25rwc\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.333269 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.333280 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.333292 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.333305 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.333317 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.333328 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.333339 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.333350 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.333360 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.333373 4739 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:31 crc kubenswrapper[4739]: I1201 15:38:31.333385 4739 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/840099a9-3ba2-4c14-bcac-de2dcf5c5d95-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 15:38:31 crc kubenswrapper[4739]: E1201 15:38:31.654599 4739 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" interval="1.6s" Dec 01 15:38:33 crc kubenswrapper[4739]: E1201 15:38:33.255799 4739 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" interval="3.2s" Dec 01 15:38:35 crc kubenswrapper[4739]: I1201 15:38:35.476984 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:35 crc kubenswrapper[4739]: I1201 15:38:35.478314 4739 status_manager.go:851] "Failed to get status for pod" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:35 crc kubenswrapper[4739]: I1201 15:38:35.478859 4739 status_manager.go:851] "Failed to get status for pod" podUID="840099a9-3ba2-4c14-bcac-de2dcf5c5d95" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-mjqqv\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:35 crc kubenswrapper[4739]: E1201 15:38:35.488965 4739 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.106:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" volumeName="registry-storage" Dec 01 15:38:35 crc kubenswrapper[4739]: I1201 15:38:35.491242 4739 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3e933a7f-4b33-4d41-bd0e-55f3ee60b933" Dec 01 15:38:35 crc kubenswrapper[4739]: I1201 15:38:35.491290 4739 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3e933a7f-4b33-4d41-bd0e-55f3ee60b933" Dec 01 15:38:35 crc kubenswrapper[4739]: E1201 15:38:35.491837 4739 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:35 crc kubenswrapper[4739]: I1201 15:38:35.492553 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:36 crc kubenswrapper[4739]: I1201 15:38:36.328332 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 15:38:36 crc kubenswrapper[4739]: I1201 15:38:36.328577 4739 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a" exitCode=1 Dec 01 15:38:36 crc kubenswrapper[4739]: I1201 15:38:36.328652 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a"} Dec 01 15:38:36 crc kubenswrapper[4739]: I1201 15:38:36.329383 4739 scope.go:117] "RemoveContainer" containerID="adf367c84048b2645ca066c407dcd3b25cbbbfdf56040d58e85c4a7d7a8b0c2a" Dec 01 15:38:36 crc kubenswrapper[4739]: I1201 15:38:36.329600 4739 status_manager.go:851] "Failed to get status for pod" podUID="840099a9-3ba2-4c14-bcac-de2dcf5c5d95" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-mjqqv\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:36 crc kubenswrapper[4739]: I1201 15:38:36.330080 4739 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:36 crc kubenswrapper[4739]: I1201 15:38:36.330499 4739 status_manager.go:851] "Failed to get status for pod" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:36 crc kubenswrapper[4739]: I1201 15:38:36.338157 4739 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="2da2a5dc2b950131ac8b062beb5f56d147b568d8cd92d1996a687004d5014fb3" exitCode=0 Dec 01 15:38:36 crc kubenswrapper[4739]: I1201 15:38:36.338216 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"2da2a5dc2b950131ac8b062beb5f56d147b568d8cd92d1996a687004d5014fb3"} Dec 01 15:38:36 crc kubenswrapper[4739]: I1201 15:38:36.338249 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"dc8c9a5f940e525465117ad3500eb7bd9f636318b7d9e593420b69509c0147ab"} Dec 01 15:38:36 crc kubenswrapper[4739]: I1201 15:38:36.338538 4739 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3e933a7f-4b33-4d41-bd0e-55f3ee60b933" Dec 01 15:38:36 crc kubenswrapper[4739]: I1201 15:38:36.338555 4739 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3e933a7f-4b33-4d41-bd0e-55f3ee60b933" Dec 01 15:38:36 crc kubenswrapper[4739]: E1201 15:38:36.339199 4739 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:36 crc kubenswrapper[4739]: I1201 15:38:36.339200 4739 status_manager.go:851] "Failed to get status for pod" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:36 crc kubenswrapper[4739]: I1201 15:38:36.339610 4739 status_manager.go:851] "Failed to get status for pod" podUID="840099a9-3ba2-4c14-bcac-de2dcf5c5d95" pod="openshift-authentication/oauth-openshift-558db77b4-mjqqv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-mjqqv\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:36 crc kubenswrapper[4739]: I1201 15:38:36.340216 4739 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:36 crc kubenswrapper[4739]: E1201 15:38:36.418957 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:38:36Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:38:36Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:38:36Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T15:38:36Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:36 crc kubenswrapper[4739]: E1201 15:38:36.419581 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:36 crc kubenswrapper[4739]: E1201 15:38:36.420097 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:36 crc kubenswrapper[4739]: E1201 15:38:36.420669 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:36 crc kubenswrapper[4739]: E1201 15:38:36.421163 4739 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" Dec 01 15:38:36 crc kubenswrapper[4739]: E1201 15:38:36.421202 4739 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 15:38:36 crc kubenswrapper[4739]: E1201 15:38:36.457272 4739 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.106:6443: connect: connection refused" interval="6.4s" Dec 01 15:38:37 crc kubenswrapper[4739]: I1201 15:38:37.355838 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 15:38:37 crc kubenswrapper[4739]: I1201 15:38:37.356203 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e328ccca69813c55431276249a578f542f6fdc661211d78530a3dbbfed707e52"} Dec 01 15:38:37 crc kubenswrapper[4739]: I1201 15:38:37.360607 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d2943213588174e9ce426a7088a600ad2d433d7224edeaeea2cf82fb6e3fcfb2"} Dec 01 15:38:37 crc kubenswrapper[4739]: I1201 15:38:37.360650 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8dcc9fd82fe9da910a4604181f94830385b499d89b656cd00d488d4b2c61886a"} Dec 01 15:38:37 crc kubenswrapper[4739]: I1201 15:38:37.360662 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5031d03afe7fd4f8d10488fa36a8d62e8e0f8ee36b18b88fa455cb3ad789eb2e"} Dec 01 15:38:37 crc kubenswrapper[4739]: I1201 15:38:37.360672 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5b7c77b7c3606b6b2a8950cf3eba1f1055519eb65032bab753eb1401b20dea68"} Dec 01 15:38:38 crc kubenswrapper[4739]: I1201 15:38:38.374529 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7f4c0cf1a1e66a7119352199aabe42c83bcf54fcbf6e70189708c01e20a88cba"} Dec 01 15:38:38 crc kubenswrapper[4739]: I1201 15:38:38.374789 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:38 crc kubenswrapper[4739]: I1201 15:38:38.374997 4739 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3e933a7f-4b33-4d41-bd0e-55f3ee60b933" Dec 01 15:38:38 crc kubenswrapper[4739]: I1201 15:38:38.375032 4739 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3e933a7f-4b33-4d41-bd0e-55f3ee60b933" Dec 01 15:38:39 crc kubenswrapper[4739]: I1201 15:38:39.622262 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:38:39 crc kubenswrapper[4739]: I1201 15:38:39.622694 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:38:39 crc kubenswrapper[4739]: I1201 15:38:39.622767 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:38:39 crc kubenswrapper[4739]: I1201 15:38:39.623635 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:38:39 crc kubenswrapper[4739]: I1201 15:38:39.623741 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e" gracePeriod=600 Dec 01 15:38:40 crc kubenswrapper[4739]: I1201 15:38:40.388636 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e" exitCode=0 Dec 01 15:38:40 crc kubenswrapper[4739]: I1201 15:38:40.388733 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e"} Dec 01 15:38:40 crc kubenswrapper[4739]: I1201 15:38:40.388974 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"682dc7ae590b7ec86c5c13fb0b344953a1f7c4e927c8f0da0655d437183e9a17"} Dec 01 15:38:40 crc kubenswrapper[4739]: I1201 15:38:40.493618 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:40 crc kubenswrapper[4739]: I1201 15:38:40.493662 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:40 crc kubenswrapper[4739]: I1201 15:38:40.498516 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:42 crc kubenswrapper[4739]: I1201 15:38:42.835485 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:38:43 crc kubenswrapper[4739]: I1201 15:38:43.405880 4739 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:44 crc kubenswrapper[4739]: I1201 15:38:44.417682 4739 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3e933a7f-4b33-4d41-bd0e-55f3ee60b933" Dec 01 15:38:44 crc kubenswrapper[4739]: I1201 15:38:44.418037 4739 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3e933a7f-4b33-4d41-bd0e-55f3ee60b933" Dec 01 15:38:44 crc kubenswrapper[4739]: I1201 15:38:44.434398 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:38:44 crc kubenswrapper[4739]: I1201 15:38:44.438034 4739 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="662c4b21-fb98-44c4-b3a0-f87eb37b4d8d" Dec 01 15:38:45 crc kubenswrapper[4739]: I1201 15:38:45.423732 4739 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3e933a7f-4b33-4d41-bd0e-55f3ee60b933" Dec 01 15:38:45 crc kubenswrapper[4739]: I1201 15:38:45.423785 4739 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3e933a7f-4b33-4d41-bd0e-55f3ee60b933" Dec 01 15:38:46 crc kubenswrapper[4739]: I1201 15:38:46.145988 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:38:46 crc kubenswrapper[4739]: I1201 15:38:46.151273 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:38:46 crc kubenswrapper[4739]: I1201 15:38:46.436755 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 15:38:48 crc kubenswrapper[4739]: I1201 15:38:48.516051 4739 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="662c4b21-fb98-44c4-b3a0-f87eb37b4d8d" Dec 01 15:38:52 crc kubenswrapper[4739]: I1201 15:38:52.510126 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 15:38:53 crc kubenswrapper[4739]: I1201 15:38:53.013219 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 15:38:53 crc kubenswrapper[4739]: I1201 15:38:53.478801 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 15:38:53 crc kubenswrapper[4739]: I1201 15:38:53.904134 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 15:38:53 crc kubenswrapper[4739]: I1201 15:38:53.992336 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 15:38:54 crc kubenswrapper[4739]: I1201 15:38:54.392560 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 15:38:54 crc kubenswrapper[4739]: I1201 15:38:54.571724 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 15:38:54 crc kubenswrapper[4739]: I1201 15:38:54.902840 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 15:38:54 crc kubenswrapper[4739]: I1201 15:38:54.910721 4739 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 15:38:55 crc kubenswrapper[4739]: I1201 15:38:55.270664 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 15:38:55 crc kubenswrapper[4739]: I1201 15:38:55.314439 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 15:38:55 crc kubenswrapper[4739]: I1201 15:38:55.380717 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 15:38:55 crc kubenswrapper[4739]: I1201 15:38:55.628077 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 15:38:55 crc kubenswrapper[4739]: I1201 15:38:55.644315 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 15:38:55 crc kubenswrapper[4739]: I1201 15:38:55.806622 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 15:38:55 crc kubenswrapper[4739]: I1201 15:38:55.887080 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 15:38:55 crc kubenswrapper[4739]: I1201 15:38:55.890387 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 15:38:55 crc kubenswrapper[4739]: I1201 15:38:55.991543 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 15:38:56 crc kubenswrapper[4739]: I1201 15:38:56.178261 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 15:38:56 crc kubenswrapper[4739]: I1201 15:38:56.477946 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 15:38:56 crc kubenswrapper[4739]: I1201 15:38:56.495501 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 15:38:56 crc kubenswrapper[4739]: I1201 15:38:56.669781 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 15:38:56 crc kubenswrapper[4739]: I1201 15:38:56.756945 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 15:38:56 crc kubenswrapper[4739]: I1201 15:38:56.916044 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 15:38:56 crc kubenswrapper[4739]: I1201 15:38:56.934614 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 15:38:56 crc kubenswrapper[4739]: I1201 15:38:56.946305 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 15:38:56 crc kubenswrapper[4739]: I1201 15:38:56.955353 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 15:38:57 crc kubenswrapper[4739]: I1201 15:38:57.287884 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 15:38:57 crc kubenswrapper[4739]: I1201 15:38:57.297188 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 15:38:57 crc kubenswrapper[4739]: I1201 15:38:57.377384 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 15:38:57 crc kubenswrapper[4739]: I1201 15:38:57.458902 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 15:38:57 crc kubenswrapper[4739]: I1201 15:38:57.678402 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 15:38:57 crc kubenswrapper[4739]: I1201 15:38:57.840997 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 15:38:57 crc kubenswrapper[4739]: I1201 15:38:57.973240 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.083350 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.092660 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.115741 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.137638 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.218590 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.228198 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.319601 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.334594 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.400019 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.421201 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.463157 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.472258 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.502980 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.552491 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.661351 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.693079 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.710065 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.740571 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.742017 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.777887 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.823868 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.863861 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 15:38:58 crc kubenswrapper[4739]: I1201 15:38:58.866584 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 15:38:59 crc kubenswrapper[4739]: I1201 15:38:59.109444 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 15:38:59 crc kubenswrapper[4739]: I1201 15:38:59.112516 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 15:38:59 crc kubenswrapper[4739]: I1201 15:38:59.179240 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 15:38:59 crc kubenswrapper[4739]: I1201 15:38:59.238899 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 15:38:59 crc kubenswrapper[4739]: I1201 15:38:59.353921 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 15:38:59 crc kubenswrapper[4739]: I1201 15:38:59.376670 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 15:38:59 crc kubenswrapper[4739]: I1201 15:38:59.417128 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 15:38:59 crc kubenswrapper[4739]: I1201 15:38:59.457152 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 15:38:59 crc kubenswrapper[4739]: I1201 15:38:59.497245 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 15:38:59 crc kubenswrapper[4739]: I1201 15:38:59.608181 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 15:38:59 crc kubenswrapper[4739]: I1201 15:38:59.669099 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 15:38:59 crc kubenswrapper[4739]: I1201 15:38:59.749087 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 15:38:59 crc kubenswrapper[4739]: I1201 15:38:59.860110 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 15:38:59 crc kubenswrapper[4739]: I1201 15:38:59.986463 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 15:38:59 crc kubenswrapper[4739]: I1201 15:38:59.994091 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.016751 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.050178 4739 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.051979 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.183752 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.471673 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.561734 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.619788 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.626165 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.667101 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.703582 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.723645 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.729195 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.783168 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.801291 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.849847 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.861723 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 15:39:00 crc kubenswrapper[4739]: I1201 15:39:00.864496 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.043156 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.045795 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.101898 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.146625 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.162357 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.202984 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.234533 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.284868 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.359141 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.467537 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.593533 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.609455 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.651193 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.655167 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.676752 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.694464 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.715704 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.816410 4739 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.817679 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.931991 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.939896 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 15:39:01 crc kubenswrapper[4739]: I1201 15:39:01.966914 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.028899 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.110896 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.176760 4739 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.178846 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.241674 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.297194 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.397289 4739 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.405473 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mjqqv","openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.405600 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.410253 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.429530 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.429507106 podStartE2EDuration="19.429507106s" podCreationTimestamp="2025-12-01 15:38:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:39:02.427273326 +0000 UTC m=+244.253019430" watchObservedRunningTime="2025-12-01 15:39:02.429507106 +0000 UTC m=+244.255253240" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.458478 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.483232 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="840099a9-3ba2-4c14-bcac-de2dcf5c5d95" path="/var/lib/kubelet/pods/840099a9-3ba2-4c14-bcac-de2dcf5c5d95/volumes" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.566217 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.576596 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.578284 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.588080 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.591107 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.652373 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.661272 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.664375 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.666965 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.693386 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.710437 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.743763 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.817018 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.872381 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.911320 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.967167 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.969440 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.974394 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 15:39:02 crc kubenswrapper[4739]: I1201 15:39:02.996810 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 15:39:03 crc kubenswrapper[4739]: I1201 15:39:03.078077 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 15:39:03 crc kubenswrapper[4739]: I1201 15:39:03.086949 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 15:39:03 crc kubenswrapper[4739]: I1201 15:39:03.146398 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 15:39:03 crc kubenswrapper[4739]: I1201 15:39:03.185473 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 15:39:03 crc kubenswrapper[4739]: I1201 15:39:03.278371 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 15:39:03 crc kubenswrapper[4739]: I1201 15:39:03.310800 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 15:39:03 crc kubenswrapper[4739]: I1201 15:39:03.316964 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 15:39:03 crc kubenswrapper[4739]: I1201 15:39:03.317386 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 15:39:03 crc kubenswrapper[4739]: I1201 15:39:03.379919 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 15:39:03 crc kubenswrapper[4739]: I1201 15:39:03.577578 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 15:39:03 crc kubenswrapper[4739]: I1201 15:39:03.698879 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 15:39:03 crc kubenswrapper[4739]: I1201 15:39:03.749770 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 15:39:03 crc kubenswrapper[4739]: I1201 15:39:03.892534 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 15:39:03 crc kubenswrapper[4739]: I1201 15:39:03.995997 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.074862 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.091610 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.102119 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.160891 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.160936 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.242554 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.256592 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.340155 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.357705 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.371493 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.436022 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.555533 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.563809 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.603577 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.630230 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.758721 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.780556 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.797937 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 15:39:04 crc kubenswrapper[4739]: I1201 15:39:04.902747 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.021069 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.036543 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.052917 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.058343 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.141303 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.276181 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.277660 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.387523 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.454887 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.597616 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.664814 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.719112 4739 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.719448 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://4161b4b7e256ffacaeceaa35d3090c913e8f2798ffd1a67318edfed1d610740f" gracePeriod=5 Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.755329 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.777159 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.799314 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.872709 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 15:39:05 crc kubenswrapper[4739]: I1201 15:39:05.929556 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.006459 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.049739 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.053103 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.113841 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.131153 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.196081 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.197872 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.199699 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.298039 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.310300 4739 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.488807 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.525133 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.595344 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.617478 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.719226 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.820503 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 15:39:06 crc kubenswrapper[4739]: I1201 15:39:06.821765 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 15:39:07 crc kubenswrapper[4739]: I1201 15:39:07.102815 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 15:39:07 crc kubenswrapper[4739]: I1201 15:39:07.215813 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 15:39:07 crc kubenswrapper[4739]: I1201 15:39:07.291863 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 15:39:07 crc kubenswrapper[4739]: I1201 15:39:07.340777 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 15:39:07 crc kubenswrapper[4739]: I1201 15:39:07.355524 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 15:39:07 crc kubenswrapper[4739]: I1201 15:39:07.618590 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 15:39:07 crc kubenswrapper[4739]: I1201 15:39:07.657580 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 15:39:07 crc kubenswrapper[4739]: I1201 15:39:07.778774 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 15:39:08 crc kubenswrapper[4739]: I1201 15:39:08.086217 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 15:39:08 crc kubenswrapper[4739]: I1201 15:39:08.212643 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 15:39:08 crc kubenswrapper[4739]: I1201 15:39:08.262200 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 15:39:08 crc kubenswrapper[4739]: I1201 15:39:08.385853 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 15:39:08 crc kubenswrapper[4739]: I1201 15:39:08.453006 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 15:39:08 crc kubenswrapper[4739]: I1201 15:39:08.559441 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 15:39:08 crc kubenswrapper[4739]: I1201 15:39:08.633745 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 15:39:08 crc kubenswrapper[4739]: I1201 15:39:08.690198 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 15:39:08 crc kubenswrapper[4739]: I1201 15:39:08.711768 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 15:39:08 crc kubenswrapper[4739]: I1201 15:39:08.861631 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 15:39:08 crc kubenswrapper[4739]: I1201 15:39:08.940071 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 15:39:09 crc kubenswrapper[4739]: I1201 15:39:09.153477 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 15:39:09 crc kubenswrapper[4739]: I1201 15:39:09.253400 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 15:39:09 crc kubenswrapper[4739]: I1201 15:39:09.263196 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 15:39:09 crc kubenswrapper[4739]: I1201 15:39:09.384036 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 15:39:09 crc kubenswrapper[4739]: I1201 15:39:09.663047 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 15:39:09 crc kubenswrapper[4739]: I1201 15:39:09.841297 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 15:39:09 crc kubenswrapper[4739]: I1201 15:39:09.854934 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 15:39:09 crc kubenswrapper[4739]: I1201 15:39:09.884040 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.125750 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.206609 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.573795 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.613465 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.877582 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.877685 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.969214 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.970990 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.971147 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.971262 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.971385 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.971478 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.971576 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.971664 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.971763 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.971852 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.972162 4739 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.972192 4739 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.972211 4739 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.972229 4739 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 15:39:10 crc kubenswrapper[4739]: I1201 15:39:10.982377 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:39:11 crc kubenswrapper[4739]: I1201 15:39:11.073840 4739 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 15:39:11 crc kubenswrapper[4739]: I1201 15:39:11.457007 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 15:39:11 crc kubenswrapper[4739]: I1201 15:39:11.581270 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 15:39:11 crc kubenswrapper[4739]: I1201 15:39:11.582145 4739 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="4161b4b7e256ffacaeceaa35d3090c913e8f2798ffd1a67318edfed1d610740f" exitCode=137 Dec 01 15:39:11 crc kubenswrapper[4739]: I1201 15:39:11.582235 4739 scope.go:117] "RemoveContainer" containerID="4161b4b7e256ffacaeceaa35d3090c913e8f2798ffd1a67318edfed1d610740f" Dec 01 15:39:11 crc kubenswrapper[4739]: I1201 15:39:11.582306 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 15:39:11 crc kubenswrapper[4739]: I1201 15:39:11.605281 4739 scope.go:117] "RemoveContainer" containerID="4161b4b7e256ffacaeceaa35d3090c913e8f2798ffd1a67318edfed1d610740f" Dec 01 15:39:11 crc kubenswrapper[4739]: E1201 15:39:11.607970 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4161b4b7e256ffacaeceaa35d3090c913e8f2798ffd1a67318edfed1d610740f\": container with ID starting with 4161b4b7e256ffacaeceaa35d3090c913e8f2798ffd1a67318edfed1d610740f not found: ID does not exist" containerID="4161b4b7e256ffacaeceaa35d3090c913e8f2798ffd1a67318edfed1d610740f" Dec 01 15:39:11 crc kubenswrapper[4739]: I1201 15:39:11.608041 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4161b4b7e256ffacaeceaa35d3090c913e8f2798ffd1a67318edfed1d610740f"} err="failed to get container status \"4161b4b7e256ffacaeceaa35d3090c913e8f2798ffd1a67318edfed1d610740f\": rpc error: code = NotFound desc = could not find container \"4161b4b7e256ffacaeceaa35d3090c913e8f2798ffd1a67318edfed1d610740f\": container with ID starting with 4161b4b7e256ffacaeceaa35d3090c913e8f2798ffd1a67318edfed1d610740f not found: ID does not exist" Dec 01 15:39:12 crc kubenswrapper[4739]: I1201 15:39:12.487621 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.266493 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-95988dd86-k9m75"] Dec 01 15:39:17 crc kubenswrapper[4739]: E1201 15:39:17.267184 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.267218 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 15:39:17 crc kubenswrapper[4739]: E1201 15:39:17.267255 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" containerName="installer" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.267271 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" containerName="installer" Dec 01 15:39:17 crc kubenswrapper[4739]: E1201 15:39:17.267306 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="840099a9-3ba2-4c14-bcac-de2dcf5c5d95" containerName="oauth-openshift" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.267319 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="840099a9-3ba2-4c14-bcac-de2dcf5c5d95" containerName="oauth-openshift" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.267509 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.267543 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="840099a9-3ba2-4c14-bcac-de2dcf5c5d95" containerName="oauth-openshift" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.267567 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="88da617e-cab3-41cc-a2d4-c6e9bd942d55" containerName="installer" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.268166 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.272999 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.273408 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.274939 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.276288 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.276682 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.278245 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.278526 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.280045 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.280600 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.281045 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.283378 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.283762 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.292030 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.335809 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-95988dd86-k9m75"] Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.344468 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.356243 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.360711 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.360789 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7c27bd13-0023-42f6-8a17-c762cc8eb539-audit-policies\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.360872 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-cliconfig\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.360965 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-user-template-login\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.361021 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-service-ca\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.361070 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-router-certs\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.361105 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-user-template-error\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.361148 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.361221 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-serving-cert\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.361291 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.361321 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-session\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.361345 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc7nw\" (UniqueName: \"kubernetes.io/projected/7c27bd13-0023-42f6-8a17-c762cc8eb539-kube-api-access-mc7nw\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.361471 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.361531 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7c27bd13-0023-42f6-8a17-c762cc8eb539-audit-dir\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.463086 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-service-ca\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.463186 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-router-certs\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.463241 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-user-template-error\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.463280 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.463331 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-serving-cert\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.463371 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.463404 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-session\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.463465 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc7nw\" (UniqueName: \"kubernetes.io/projected/7c27bd13-0023-42f6-8a17-c762cc8eb539-kube-api-access-mc7nw\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.463502 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.463541 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7c27bd13-0023-42f6-8a17-c762cc8eb539-audit-dir\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.463598 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.463644 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7c27bd13-0023-42f6-8a17-c762cc8eb539-audit-policies\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.463694 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-cliconfig\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.463735 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-user-template-login\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.464064 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7c27bd13-0023-42f6-8a17-c762cc8eb539-audit-dir\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.465398 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-service-ca\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.465833 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7c27bd13-0023-42f6-8a17-c762cc8eb539-audit-policies\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.466258 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-cliconfig\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.466723 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.470043 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.470147 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-router-certs\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.470222 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-session\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.471110 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-user-template-error\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.473379 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.474259 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-system-serving-cert\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.474464 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-user-template-login\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.474814 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7c27bd13-0023-42f6-8a17-c762cc8eb539-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.494026 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc7nw\" (UniqueName: \"kubernetes.io/projected/7c27bd13-0023-42f6-8a17-c762cc8eb539-kube-api-access-mc7nw\") pod \"oauth-openshift-95988dd86-k9m75\" (UID: \"7c27bd13-0023-42f6-8a17-c762cc8eb539\") " pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.648562 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:17 crc kubenswrapper[4739]: I1201 15:39:17.923037 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-95988dd86-k9m75"] Dec 01 15:39:18 crc kubenswrapper[4739]: I1201 15:39:18.631455 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" event={"ID":"7c27bd13-0023-42f6-8a17-c762cc8eb539","Type":"ContainerStarted","Data":"d7eb31ab6c26147f37dddd63ad694086b5311b97cc0f41460a0dc10258a03e24"} Dec 01 15:39:18 crc kubenswrapper[4739]: I1201 15:39:18.631849 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:18 crc kubenswrapper[4739]: I1201 15:39:18.631870 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" event={"ID":"7c27bd13-0023-42f6-8a17-c762cc8eb539","Type":"ContainerStarted","Data":"570aa0ff4699017ac55d9fe1633d83a7b64f341dc50844aeff1d84b7f2383217"} Dec 01 15:39:18 crc kubenswrapper[4739]: I1201 15:39:18.738388 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" Dec 01 15:39:18 crc kubenswrapper[4739]: I1201 15:39:18.761285 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-95988dd86-k9m75" podStartSLOduration=73.761268187 podStartE2EDuration="1m13.761268187s" podCreationTimestamp="2025-12-01 15:38:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:39:18.707453301 +0000 UTC m=+260.533199395" watchObservedRunningTime="2025-12-01 15:39:18.761268187 +0000 UTC m=+260.587014291" Dec 01 15:39:30 crc kubenswrapper[4739]: I1201 15:39:30.714401 4739 generic.go:334] "Generic (PLEG): container finished" podID="9cbb8f14-1426-4c22-8ccb-f81dbdc8f175" containerID="c17daf213ad6ae6b3ce703966eaf27983cec45d5eff97818ffb41d64f349f794" exitCode=0 Dec 01 15:39:30 crc kubenswrapper[4739]: I1201 15:39:30.715052 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" event={"ID":"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175","Type":"ContainerDied","Data":"c17daf213ad6ae6b3ce703966eaf27983cec45d5eff97818ffb41d64f349f794"} Dec 01 15:39:30 crc kubenswrapper[4739]: I1201 15:39:30.715694 4739 scope.go:117] "RemoveContainer" containerID="c17daf213ad6ae6b3ce703966eaf27983cec45d5eff97818ffb41d64f349f794" Dec 01 15:39:31 crc kubenswrapper[4739]: I1201 15:39:31.723584 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" event={"ID":"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175","Type":"ContainerStarted","Data":"38ce4199e7ed95d416cb9e34b0d1992d0babbec85ec2f49c84bf4a889e54ed6d"} Dec 01 15:39:31 crc kubenswrapper[4739]: I1201 15:39:31.724557 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:39:31 crc kubenswrapper[4739]: I1201 15:39:31.731151 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:39:47 crc kubenswrapper[4739]: I1201 15:39:47.724652 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-b9xb8"] Dec 01 15:39:47 crc kubenswrapper[4739]: I1201 15:39:47.725477 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" podUID="6cec3a69-7dbc-4781-a800-452a7e740adc" containerName="controller-manager" containerID="cri-o://e29a251c5743ed9b8763ff2af19e9f7b99969df8f44297d1db12abc7fbecb33b" gracePeriod=30 Dec 01 15:39:47 crc kubenswrapper[4739]: I1201 15:39:47.817091 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5"] Dec 01 15:39:47 crc kubenswrapper[4739]: I1201 15:39:47.817336 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" podUID="c886dc3f-4a59-4e31-a900-2c000427c7a9" containerName="route-controller-manager" containerID="cri-o://aec86c2c39bf8605c5471a9dda6f5f4e0828731916faf83a29386ff62e7499db" gracePeriod=30 Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.104962 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.142913 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.249599 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cec3a69-7dbc-4781-a800-452a7e740adc-serving-cert\") pod \"6cec3a69-7dbc-4781-a800-452a7e740adc\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.249706 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-client-ca\") pod \"6cec3a69-7dbc-4781-a800-452a7e740adc\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.249791 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-config\") pod \"6cec3a69-7dbc-4781-a800-452a7e740adc\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.249841 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcl6r\" (UniqueName: \"kubernetes.io/projected/6cec3a69-7dbc-4781-a800-452a7e740adc-kube-api-access-mcl6r\") pod \"6cec3a69-7dbc-4781-a800-452a7e740adc\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.249888 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-proxy-ca-bundles\") pod \"6cec3a69-7dbc-4781-a800-452a7e740adc\" (UID: \"6cec3a69-7dbc-4781-a800-452a7e740adc\") " Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.249920 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c886dc3f-4a59-4e31-a900-2c000427c7a9-client-ca\") pod \"c886dc3f-4a59-4e31-a900-2c000427c7a9\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.249956 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9r6cx\" (UniqueName: \"kubernetes.io/projected/c886dc3f-4a59-4e31-a900-2c000427c7a9-kube-api-access-9r6cx\") pod \"c886dc3f-4a59-4e31-a900-2c000427c7a9\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.249994 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c886dc3f-4a59-4e31-a900-2c000427c7a9-serving-cert\") pod \"c886dc3f-4a59-4e31-a900-2c000427c7a9\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.250019 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c886dc3f-4a59-4e31-a900-2c000427c7a9-config\") pod \"c886dc3f-4a59-4e31-a900-2c000427c7a9\" (UID: \"c886dc3f-4a59-4e31-a900-2c000427c7a9\") " Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.250935 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-client-ca" (OuterVolumeSpecName: "client-ca") pod "6cec3a69-7dbc-4781-a800-452a7e740adc" (UID: "6cec3a69-7dbc-4781-a800-452a7e740adc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.251056 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "6cec3a69-7dbc-4781-a800-452a7e740adc" (UID: "6cec3a69-7dbc-4781-a800-452a7e740adc"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.251300 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c886dc3f-4a59-4e31-a900-2c000427c7a9-config" (OuterVolumeSpecName: "config") pod "c886dc3f-4a59-4e31-a900-2c000427c7a9" (UID: "c886dc3f-4a59-4e31-a900-2c000427c7a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.252146 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-config" (OuterVolumeSpecName: "config") pod "6cec3a69-7dbc-4781-a800-452a7e740adc" (UID: "6cec3a69-7dbc-4781-a800-452a7e740adc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.253089 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c886dc3f-4a59-4e31-a900-2c000427c7a9-client-ca" (OuterVolumeSpecName: "client-ca") pod "c886dc3f-4a59-4e31-a900-2c000427c7a9" (UID: "c886dc3f-4a59-4e31-a900-2c000427c7a9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.255942 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cec3a69-7dbc-4781-a800-452a7e740adc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6cec3a69-7dbc-4781-a800-452a7e740adc" (UID: "6cec3a69-7dbc-4781-a800-452a7e740adc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.256276 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cec3a69-7dbc-4781-a800-452a7e740adc-kube-api-access-mcl6r" (OuterVolumeSpecName: "kube-api-access-mcl6r") pod "6cec3a69-7dbc-4781-a800-452a7e740adc" (UID: "6cec3a69-7dbc-4781-a800-452a7e740adc"). InnerVolumeSpecName "kube-api-access-mcl6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.256860 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c886dc3f-4a59-4e31-a900-2c000427c7a9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c886dc3f-4a59-4e31-a900-2c000427c7a9" (UID: "c886dc3f-4a59-4e31-a900-2c000427c7a9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.258113 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c886dc3f-4a59-4e31-a900-2c000427c7a9-kube-api-access-9r6cx" (OuterVolumeSpecName: "kube-api-access-9r6cx") pod "c886dc3f-4a59-4e31-a900-2c000427c7a9" (UID: "c886dc3f-4a59-4e31-a900-2c000427c7a9"). InnerVolumeSpecName "kube-api-access-9r6cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.352190 4739 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.352265 4739 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c886dc3f-4a59-4e31-a900-2c000427c7a9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.352293 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9r6cx\" (UniqueName: \"kubernetes.io/projected/c886dc3f-4a59-4e31-a900-2c000427c7a9-kube-api-access-9r6cx\") on node \"crc\" DevicePath \"\"" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.352322 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c886dc3f-4a59-4e31-a900-2c000427c7a9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.352346 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c886dc3f-4a59-4e31-a900-2c000427c7a9-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.352370 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cec3a69-7dbc-4781-a800-452a7e740adc-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.352392 4739 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.352414 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cec3a69-7dbc-4781-a800-452a7e740adc-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.352470 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcl6r\" (UniqueName: \"kubernetes.io/projected/6cec3a69-7dbc-4781-a800-452a7e740adc-kube-api-access-mcl6r\") on node \"crc\" DevicePath \"\"" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.827055 4739 generic.go:334] "Generic (PLEG): container finished" podID="6cec3a69-7dbc-4781-a800-452a7e740adc" containerID="e29a251c5743ed9b8763ff2af19e9f7b99969df8f44297d1db12abc7fbecb33b" exitCode=0 Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.827143 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.827166 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" event={"ID":"6cec3a69-7dbc-4781-a800-452a7e740adc","Type":"ContainerDied","Data":"e29a251c5743ed9b8763ff2af19e9f7b99969df8f44297d1db12abc7fbecb33b"} Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.827259 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-b9xb8" event={"ID":"6cec3a69-7dbc-4781-a800-452a7e740adc","Type":"ContainerDied","Data":"f3904bcc1368a32b4ba00c515cd23cf4c5614484b366eda221c7a05b8437cc59"} Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.827320 4739 scope.go:117] "RemoveContainer" containerID="e29a251c5743ed9b8763ff2af19e9f7b99969df8f44297d1db12abc7fbecb33b" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.833745 4739 generic.go:334] "Generic (PLEG): container finished" podID="c886dc3f-4a59-4e31-a900-2c000427c7a9" containerID="aec86c2c39bf8605c5471a9dda6f5f4e0828731916faf83a29386ff62e7499db" exitCode=0 Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.833789 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.833808 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" event={"ID":"c886dc3f-4a59-4e31-a900-2c000427c7a9","Type":"ContainerDied","Data":"aec86c2c39bf8605c5471a9dda6f5f4e0828731916faf83a29386ff62e7499db"} Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.833848 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5" event={"ID":"c886dc3f-4a59-4e31-a900-2c000427c7a9","Type":"ContainerDied","Data":"9e96c9f14e76bed19291d59816a8b5cdacaa237f09ee9a8a127eec2a381df6c8"} Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.879656 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-b9xb8"] Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.887436 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-b9xb8"] Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.887461 4739 scope.go:117] "RemoveContainer" containerID="e29a251c5743ed9b8763ff2af19e9f7b99969df8f44297d1db12abc7fbecb33b" Dec 01 15:39:48 crc kubenswrapper[4739]: E1201 15:39:48.888242 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e29a251c5743ed9b8763ff2af19e9f7b99969df8f44297d1db12abc7fbecb33b\": container with ID starting with e29a251c5743ed9b8763ff2af19e9f7b99969df8f44297d1db12abc7fbecb33b not found: ID does not exist" containerID="e29a251c5743ed9b8763ff2af19e9f7b99969df8f44297d1db12abc7fbecb33b" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.888291 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e29a251c5743ed9b8763ff2af19e9f7b99969df8f44297d1db12abc7fbecb33b"} err="failed to get container status \"e29a251c5743ed9b8763ff2af19e9f7b99969df8f44297d1db12abc7fbecb33b\": rpc error: code = NotFound desc = could not find container \"e29a251c5743ed9b8763ff2af19e9f7b99969df8f44297d1db12abc7fbecb33b\": container with ID starting with e29a251c5743ed9b8763ff2af19e9f7b99969df8f44297d1db12abc7fbecb33b not found: ID does not exist" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.888321 4739 scope.go:117] "RemoveContainer" containerID="aec86c2c39bf8605c5471a9dda6f5f4e0828731916faf83a29386ff62e7499db" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.895393 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5"] Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.898942 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5pnp5"] Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.913737 4739 scope.go:117] "RemoveContainer" containerID="aec86c2c39bf8605c5471a9dda6f5f4e0828731916faf83a29386ff62e7499db" Dec 01 15:39:48 crc kubenswrapper[4739]: E1201 15:39:48.914275 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aec86c2c39bf8605c5471a9dda6f5f4e0828731916faf83a29386ff62e7499db\": container with ID starting with aec86c2c39bf8605c5471a9dda6f5f4e0828731916faf83a29386ff62e7499db not found: ID does not exist" containerID="aec86c2c39bf8605c5471a9dda6f5f4e0828731916faf83a29386ff62e7499db" Dec 01 15:39:48 crc kubenswrapper[4739]: I1201 15:39:48.914313 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aec86c2c39bf8605c5471a9dda6f5f4e0828731916faf83a29386ff62e7499db"} err="failed to get container status \"aec86c2c39bf8605c5471a9dda6f5f4e0828731916faf83a29386ff62e7499db\": rpc error: code = NotFound desc = could not find container \"aec86c2c39bf8605c5471a9dda6f5f4e0828731916faf83a29386ff62e7499db\": container with ID starting with aec86c2c39bf8605c5471a9dda6f5f4e0828731916faf83a29386ff62e7499db not found: ID does not exist" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.285742 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz"] Dec 01 15:39:49 crc kubenswrapper[4739]: E1201 15:39:49.286138 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cec3a69-7dbc-4781-a800-452a7e740adc" containerName="controller-manager" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.286167 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cec3a69-7dbc-4781-a800-452a7e740adc" containerName="controller-manager" Dec 01 15:39:49 crc kubenswrapper[4739]: E1201 15:39:49.286184 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c886dc3f-4a59-4e31-a900-2c000427c7a9" containerName="route-controller-manager" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.286197 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="c886dc3f-4a59-4e31-a900-2c000427c7a9" containerName="route-controller-manager" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.286390 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cec3a69-7dbc-4781-a800-452a7e740adc" containerName="controller-manager" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.286443 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="c886dc3f-4a59-4e31-a900-2c000427c7a9" containerName="route-controller-manager" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.287151 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.294203 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-bf64b4967-gbbbn"] Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.295300 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.299929 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.299996 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.300409 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.300453 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.301279 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.301434 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.301553 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.301721 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.303329 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.303618 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.304569 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.304945 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.310759 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz"] Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.315608 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-bf64b4967-gbbbn"] Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.317911 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.367662 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20f7f236-e00a-4002-9056-5099d2a6e485-serving-cert\") pod \"controller-manager-bf64b4967-gbbbn\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.367705 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82mgd\" (UniqueName: \"kubernetes.io/projected/192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b-kube-api-access-82mgd\") pod \"route-controller-manager-697f44cb6c-hnjzz\" (UID: \"192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.367749 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b-config\") pod \"route-controller-manager-697f44cb6c-hnjzz\" (UID: \"192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.367775 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9cch\" (UniqueName: \"kubernetes.io/projected/20f7f236-e00a-4002-9056-5099d2a6e485-kube-api-access-z9cch\") pod \"controller-manager-bf64b4967-gbbbn\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.367820 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b-serving-cert\") pod \"route-controller-manager-697f44cb6c-hnjzz\" (UID: \"192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.367863 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b-client-ca\") pod \"route-controller-manager-697f44cb6c-hnjzz\" (UID: \"192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.367941 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-client-ca\") pod \"controller-manager-bf64b4967-gbbbn\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.367967 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-proxy-ca-bundles\") pod \"controller-manager-bf64b4967-gbbbn\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.368011 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-config\") pod \"controller-manager-bf64b4967-gbbbn\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.469169 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b-client-ca\") pod \"route-controller-manager-697f44cb6c-hnjzz\" (UID: \"192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.469261 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-client-ca\") pod \"controller-manager-bf64b4967-gbbbn\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.469299 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-proxy-ca-bundles\") pod \"controller-manager-bf64b4967-gbbbn\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.469363 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-config\") pod \"controller-manager-bf64b4967-gbbbn\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.469403 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20f7f236-e00a-4002-9056-5099d2a6e485-serving-cert\") pod \"controller-manager-bf64b4967-gbbbn\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.469463 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82mgd\" (UniqueName: \"kubernetes.io/projected/192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b-kube-api-access-82mgd\") pod \"route-controller-manager-697f44cb6c-hnjzz\" (UID: \"192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.469512 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b-config\") pod \"route-controller-manager-697f44cb6c-hnjzz\" (UID: \"192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.469547 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9cch\" (UniqueName: \"kubernetes.io/projected/20f7f236-e00a-4002-9056-5099d2a6e485-kube-api-access-z9cch\") pod \"controller-manager-bf64b4967-gbbbn\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.469595 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b-serving-cert\") pod \"route-controller-manager-697f44cb6c-hnjzz\" (UID: \"192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.471073 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b-client-ca\") pod \"route-controller-manager-697f44cb6c-hnjzz\" (UID: \"192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.471087 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-client-ca\") pod \"controller-manager-bf64b4967-gbbbn\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.471647 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b-config\") pod \"route-controller-manager-697f44cb6c-hnjzz\" (UID: \"192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.471748 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-proxy-ca-bundles\") pod \"controller-manager-bf64b4967-gbbbn\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.471934 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-config\") pod \"controller-manager-bf64b4967-gbbbn\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.475207 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b-serving-cert\") pod \"route-controller-manager-697f44cb6c-hnjzz\" (UID: \"192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.476062 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20f7f236-e00a-4002-9056-5099d2a6e485-serving-cert\") pod \"controller-manager-bf64b4967-gbbbn\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.502319 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9cch\" (UniqueName: \"kubernetes.io/projected/20f7f236-e00a-4002-9056-5099d2a6e485-kube-api-access-z9cch\") pod \"controller-manager-bf64b4967-gbbbn\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.502375 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82mgd\" (UniqueName: \"kubernetes.io/projected/192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b-kube-api-access-82mgd\") pod \"route-controller-manager-697f44cb6c-hnjzz\" (UID: \"192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.623544 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:39:49 crc kubenswrapper[4739]: I1201 15:39:49.634889 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:50 crc kubenswrapper[4739]: I1201 15:39:50.271827 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-bf64b4967-gbbbn"] Dec 01 15:39:50 crc kubenswrapper[4739]: W1201 15:39:50.276197 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20f7f236_e00a_4002_9056_5099d2a6e485.slice/crio-b486cb9f887625bc111ae3cf1e3ff7664aa89c6706420a86493da17c43553c3e WatchSource:0}: Error finding container b486cb9f887625bc111ae3cf1e3ff7664aa89c6706420a86493da17c43553c3e: Status 404 returned error can't find the container with id b486cb9f887625bc111ae3cf1e3ff7664aa89c6706420a86493da17c43553c3e Dec 01 15:39:50 crc kubenswrapper[4739]: I1201 15:39:50.312933 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz"] Dec 01 15:39:50 crc kubenswrapper[4739]: W1201 15:39:50.315611 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod192d74ed_cea5_4fc5_88b0_f26ccb6f9f8b.slice/crio-b6d9845100c79165d925faf10508e021d959da83d94671fa6027deb6ab0ca9e4 WatchSource:0}: Error finding container b6d9845100c79165d925faf10508e021d959da83d94671fa6027deb6ab0ca9e4: Status 404 returned error can't find the container with id b6d9845100c79165d925faf10508e021d959da83d94671fa6027deb6ab0ca9e4 Dec 01 15:39:50 crc kubenswrapper[4739]: I1201 15:39:50.485804 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cec3a69-7dbc-4781-a800-452a7e740adc" path="/var/lib/kubelet/pods/6cec3a69-7dbc-4781-a800-452a7e740adc/volumes" Dec 01 15:39:50 crc kubenswrapper[4739]: I1201 15:39:50.486467 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c886dc3f-4a59-4e31-a900-2c000427c7a9" path="/var/lib/kubelet/pods/c886dc3f-4a59-4e31-a900-2c000427c7a9/volumes" Dec 01 15:39:51 crc kubenswrapper[4739]: I1201 15:39:51.129549 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" event={"ID":"192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b","Type":"ContainerStarted","Data":"fdb96c235967d12ca6440e682dda0b5218709cb53b78f12f1c81a521b0146bd8"} Dec 01 15:39:51 crc kubenswrapper[4739]: I1201 15:39:51.129596 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" event={"ID":"192d74ed-cea5-4fc5-88b0-f26ccb6f9f8b","Type":"ContainerStarted","Data":"b6d9845100c79165d925faf10508e021d959da83d94671fa6027deb6ab0ca9e4"} Dec 01 15:39:51 crc kubenswrapper[4739]: I1201 15:39:51.131558 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:39:51 crc kubenswrapper[4739]: I1201 15:39:51.133152 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" event={"ID":"20f7f236-e00a-4002-9056-5099d2a6e485","Type":"ContainerStarted","Data":"3cb290daeb34ddfc00da0252d0c45471ff170aa7e2971968d423e25871284cdb"} Dec 01 15:39:51 crc kubenswrapper[4739]: I1201 15:39:51.133194 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" event={"ID":"20f7f236-e00a-4002-9056-5099d2a6e485","Type":"ContainerStarted","Data":"b486cb9f887625bc111ae3cf1e3ff7664aa89c6706420a86493da17c43553c3e"} Dec 01 15:39:51 crc kubenswrapper[4739]: I1201 15:39:51.133811 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:51 crc kubenswrapper[4739]: I1201 15:39:51.137359 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:39:51 crc kubenswrapper[4739]: I1201 15:39:51.153481 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" podStartSLOduration=4.153461937 podStartE2EDuration="4.153461937s" podCreationTimestamp="2025-12-01 15:39:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:39:51.148467476 +0000 UTC m=+292.974213580" watchObservedRunningTime="2025-12-01 15:39:51.153461937 +0000 UTC m=+292.979208041" Dec 01 15:39:51 crc kubenswrapper[4739]: I1201 15:39:51.166469 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" podStartSLOduration=4.166452573 podStartE2EDuration="4.166452573s" podCreationTimestamp="2025-12-01 15:39:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:39:51.165980688 +0000 UTC m=+292.991726792" watchObservedRunningTime="2025-12-01 15:39:51.166452573 +0000 UTC m=+292.992198677" Dec 01 15:39:51 crc kubenswrapper[4739]: I1201 15:39:51.175477 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-hnjzz" Dec 01 15:40:06 crc kubenswrapper[4739]: I1201 15:40:06.952321 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8vzjc"] Dec 01 15:40:06 crc kubenswrapper[4739]: I1201 15:40:06.953301 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8vzjc" podUID="cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" containerName="registry-server" containerID="cri-o://69ad0151a96195be86ef2057a7e224dabf750cdfcfaf752078ec3f058613a84b" gracePeriod=30 Dec 01 15:40:06 crc kubenswrapper[4739]: I1201 15:40:06.955621 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwqtt"] Dec 01 15:40:06 crc kubenswrapper[4739]: I1201 15:40:06.955847 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nwqtt" podUID="5b9930dd-3336-438d-84fc-3a62b6a86e75" containerName="registry-server" containerID="cri-o://72f42d0ff09f68b5d8669e5dc48d5fef6bae5276fd9dc82aee41e25ce11854e9" gracePeriod=30 Dec 01 15:40:06 crc kubenswrapper[4739]: I1201 15:40:06.969791 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l7mfx"] Dec 01 15:40:06 crc kubenswrapper[4739]: I1201 15:40:06.970116 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" podUID="9cbb8f14-1426-4c22-8ccb-f81dbdc8f175" containerName="marketplace-operator" containerID="cri-o://38ce4199e7ed95d416cb9e34b0d1992d0babbec85ec2f49c84bf4a889e54ed6d" gracePeriod=30 Dec 01 15:40:06 crc kubenswrapper[4739]: I1201 15:40:06.988887 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z5d6l"] Dec 01 15:40:06 crc kubenswrapper[4739]: I1201 15:40:06.989164 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z5d6l" podUID="44224079-e66e-4e6f-a74e-5595d3a378f2" containerName="registry-server" containerID="cri-o://617ee75a624932ed3767f42d74925f8a0e7f65dd4132310fbdb8d4fdb6607b8f" gracePeriod=30 Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.004576 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8mjpt"] Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.010803 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.012219 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ksr6g"] Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.015246 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8mjpt"] Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.082597 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a480b7ca-0552-40d0-ae91-cccfcba31229-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8mjpt\" (UID: \"a480b7ca-0552-40d0-ae91-cccfcba31229\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.082662 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a480b7ca-0552-40d0-ae91-cccfcba31229-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8mjpt\" (UID: \"a480b7ca-0552-40d0-ae91-cccfcba31229\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.082746 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfm8w\" (UniqueName: \"kubernetes.io/projected/a480b7ca-0552-40d0-ae91-cccfcba31229-kube-api-access-qfm8w\") pod \"marketplace-operator-79b997595-8mjpt\" (UID: \"a480b7ca-0552-40d0-ae91-cccfcba31229\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.183890 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a480b7ca-0552-40d0-ae91-cccfcba31229-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8mjpt\" (UID: \"a480b7ca-0552-40d0-ae91-cccfcba31229\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.184239 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a480b7ca-0552-40d0-ae91-cccfcba31229-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8mjpt\" (UID: \"a480b7ca-0552-40d0-ae91-cccfcba31229\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.184275 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfm8w\" (UniqueName: \"kubernetes.io/projected/a480b7ca-0552-40d0-ae91-cccfcba31229-kube-api-access-qfm8w\") pod \"marketplace-operator-79b997595-8mjpt\" (UID: \"a480b7ca-0552-40d0-ae91-cccfcba31229\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.185866 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a480b7ca-0552-40d0-ae91-cccfcba31229-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8mjpt\" (UID: \"a480b7ca-0552-40d0-ae91-cccfcba31229\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.188961 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a480b7ca-0552-40d0-ae91-cccfcba31229-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8mjpt\" (UID: \"a480b7ca-0552-40d0-ae91-cccfcba31229\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.214803 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfm8w\" (UniqueName: \"kubernetes.io/projected/a480b7ca-0552-40d0-ae91-cccfcba31229-kube-api-access-qfm8w\") pod \"marketplace-operator-79b997595-8mjpt\" (UID: \"a480b7ca-0552-40d0-ae91-cccfcba31229\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.227919 4739 generic.go:334] "Generic (PLEG): container finished" podID="9cbb8f14-1426-4c22-8ccb-f81dbdc8f175" containerID="38ce4199e7ed95d416cb9e34b0d1992d0babbec85ec2f49c84bf4a889e54ed6d" exitCode=0 Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.228065 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" event={"ID":"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175","Type":"ContainerDied","Data":"38ce4199e7ed95d416cb9e34b0d1992d0babbec85ec2f49c84bf4a889e54ed6d"} Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.228162 4739 scope.go:117] "RemoveContainer" containerID="c17daf213ad6ae6b3ce703966eaf27983cec45d5eff97818ffb41d64f349f794" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.230695 4739 generic.go:334] "Generic (PLEG): container finished" podID="cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" containerID="69ad0151a96195be86ef2057a7e224dabf750cdfcfaf752078ec3f058613a84b" exitCode=0 Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.230759 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vzjc" event={"ID":"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4","Type":"ContainerDied","Data":"69ad0151a96195be86ef2057a7e224dabf750cdfcfaf752078ec3f058613a84b"} Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.232887 4739 generic.go:334] "Generic (PLEG): container finished" podID="44224079-e66e-4e6f-a74e-5595d3a378f2" containerID="617ee75a624932ed3767f42d74925f8a0e7f65dd4132310fbdb8d4fdb6607b8f" exitCode=0 Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.232966 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5d6l" event={"ID":"44224079-e66e-4e6f-a74e-5595d3a378f2","Type":"ContainerDied","Data":"617ee75a624932ed3767f42d74925f8a0e7f65dd4132310fbdb8d4fdb6607b8f"} Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.234618 4739 generic.go:334] "Generic (PLEG): container finished" podID="5b9930dd-3336-438d-84fc-3a62b6a86e75" containerID="72f42d0ff09f68b5d8669e5dc48d5fef6bae5276fd9dc82aee41e25ce11854e9" exitCode=0 Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.234705 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwqtt" event={"ID":"5b9930dd-3336-438d-84fc-3a62b6a86e75","Type":"ContainerDied","Data":"72f42d0ff09f68b5d8669e5dc48d5fef6bae5276fd9dc82aee41e25ce11854e9"} Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.234931 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ksr6g" podUID="83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8" containerName="registry-server" containerID="cri-o://ff70135236a4e596926d23d394ea708c66ab95f21964efd0cb13000422d59f74" gracePeriod=30 Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.272670 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.475635 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.491347 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b9930dd-3336-438d-84fc-3a62b6a86e75-catalog-content\") pod \"5b9930dd-3336-438d-84fc-3a62b6a86e75\" (UID: \"5b9930dd-3336-438d-84fc-3a62b6a86e75\") " Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.491564 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b9930dd-3336-438d-84fc-3a62b6a86e75-utilities\") pod \"5b9930dd-3336-438d-84fc-3a62b6a86e75\" (UID: \"5b9930dd-3336-438d-84fc-3a62b6a86e75\") " Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.492192 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlwr9\" (UniqueName: \"kubernetes.io/projected/5b9930dd-3336-438d-84fc-3a62b6a86e75-kube-api-access-zlwr9\") pod \"5b9930dd-3336-438d-84fc-3a62b6a86e75\" (UID: \"5b9930dd-3336-438d-84fc-3a62b6a86e75\") " Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.500807 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b9930dd-3336-438d-84fc-3a62b6a86e75-utilities" (OuterVolumeSpecName: "utilities") pod "5b9930dd-3336-438d-84fc-3a62b6a86e75" (UID: "5b9930dd-3336-438d-84fc-3a62b6a86e75"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.507169 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b9930dd-3336-438d-84fc-3a62b6a86e75-kube-api-access-zlwr9" (OuterVolumeSpecName: "kube-api-access-zlwr9") pod "5b9930dd-3336-438d-84fc-3a62b6a86e75" (UID: "5b9930dd-3336-438d-84fc-3a62b6a86e75"). InnerVolumeSpecName "kube-api-access-zlwr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.553951 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b9930dd-3336-438d-84fc-3a62b6a86e75-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b9930dd-3336-438d-84fc-3a62b6a86e75" (UID: "5b9930dd-3336-438d-84fc-3a62b6a86e75"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.593544 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlwr9\" (UniqueName: \"kubernetes.io/projected/5b9930dd-3336-438d-84fc-3a62b6a86e75-kube-api-access-zlwr9\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.593584 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b9930dd-3336-438d-84fc-3a62b6a86e75-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.593593 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b9930dd-3336-438d-84fc-3a62b6a86e75-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.601038 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.620884 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.637353 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.694302 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44224079-e66e-4e6f-a74e-5595d3a378f2-utilities\") pod \"44224079-e66e-4e6f-a74e-5595d3a378f2\" (UID: \"44224079-e66e-4e6f-a74e-5595d3a378f2\") " Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.694347 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-utilities\") pod \"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4\" (UID: \"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4\") " Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.694400 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbp29\" (UniqueName: \"kubernetes.io/projected/44224079-e66e-4e6f-a74e-5595d3a378f2-kube-api-access-jbp29\") pod \"44224079-e66e-4e6f-a74e-5595d3a378f2\" (UID: \"44224079-e66e-4e6f-a74e-5595d3a378f2\") " Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.694440 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75fd8\" (UniqueName: \"kubernetes.io/projected/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-kube-api-access-75fd8\") pod \"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4\" (UID: \"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4\") " Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.694462 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-marketplace-operator-metrics\") pod \"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175\" (UID: \"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175\") " Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.694527 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-marketplace-trusted-ca\") pod \"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175\" (UID: \"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175\") " Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.694547 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vklx7\" (UniqueName: \"kubernetes.io/projected/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-kube-api-access-vklx7\") pod \"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175\" (UID: \"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175\") " Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.694574 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-catalog-content\") pod \"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4\" (UID: \"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4\") " Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.694600 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44224079-e66e-4e6f-a74e-5595d3a378f2-catalog-content\") pod \"44224079-e66e-4e6f-a74e-5595d3a378f2\" (UID: \"44224079-e66e-4e6f-a74e-5595d3a378f2\") " Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.695088 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44224079-e66e-4e6f-a74e-5595d3a378f2-utilities" (OuterVolumeSpecName: "utilities") pod "44224079-e66e-4e6f-a74e-5595d3a378f2" (UID: "44224079-e66e-4e6f-a74e-5595d3a378f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.695262 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "9cbb8f14-1426-4c22-8ccb-f81dbdc8f175" (UID: "9cbb8f14-1426-4c22-8ccb-f81dbdc8f175"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.695952 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-utilities" (OuterVolumeSpecName: "utilities") pod "cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" (UID: "cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.698060 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "9cbb8f14-1426-4c22-8ccb-f81dbdc8f175" (UID: "9cbb8f14-1426-4c22-8ccb-f81dbdc8f175"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.698401 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-kube-api-access-vklx7" (OuterVolumeSpecName: "kube-api-access-vklx7") pod "9cbb8f14-1426-4c22-8ccb-f81dbdc8f175" (UID: "9cbb8f14-1426-4c22-8ccb-f81dbdc8f175"). InnerVolumeSpecName "kube-api-access-vklx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.700674 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-kube-api-access-75fd8" (OuterVolumeSpecName: "kube-api-access-75fd8") pod "cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" (UID: "cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4"). InnerVolumeSpecName "kube-api-access-75fd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.719798 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44224079-e66e-4e6f-a74e-5595d3a378f2-kube-api-access-jbp29" (OuterVolumeSpecName: "kube-api-access-jbp29") pod "44224079-e66e-4e6f-a74e-5595d3a378f2" (UID: "44224079-e66e-4e6f-a74e-5595d3a378f2"). InnerVolumeSpecName "kube-api-access-jbp29". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.722053 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44224079-e66e-4e6f-a74e-5595d3a378f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44224079-e66e-4e6f-a74e-5595d3a378f2" (UID: "44224079-e66e-4e6f-a74e-5595d3a378f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.725112 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.737072 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" (UID: "cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.795663 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-utilities\") pod \"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8\" (UID: \"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8\") " Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.795713 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-catalog-content\") pod \"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8\" (UID: \"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8\") " Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.795742 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4s5b\" (UniqueName: \"kubernetes.io/projected/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-kube-api-access-c4s5b\") pod \"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8\" (UID: \"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8\") " Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.795892 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbp29\" (UniqueName: \"kubernetes.io/projected/44224079-e66e-4e6f-a74e-5595d3a378f2-kube-api-access-jbp29\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.795904 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75fd8\" (UniqueName: \"kubernetes.io/projected/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-kube-api-access-75fd8\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.795913 4739 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.795921 4739 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.795929 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vklx7\" (UniqueName: \"kubernetes.io/projected/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175-kube-api-access-vklx7\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.795938 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.795948 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44224079-e66e-4e6f-a74e-5595d3a378f2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.795956 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44224079-e66e-4e6f-a74e-5595d3a378f2-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.795964 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.796894 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-utilities" (OuterVolumeSpecName: "utilities") pod "83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8" (UID: "83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.797998 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-kube-api-access-c4s5b" (OuterVolumeSpecName: "kube-api-access-c4s5b") pod "83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8" (UID: "83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8"). InnerVolumeSpecName "kube-api-access-c4s5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.862271 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8mjpt"] Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.892974 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8" (UID: "83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.896959 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.896987 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4s5b\" (UniqueName: \"kubernetes.io/projected/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-kube-api-access-c4s5b\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:07 crc kubenswrapper[4739]: I1201 15:40:07.896999 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.240730 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" event={"ID":"9cbb8f14-1426-4c22-8ccb-f81dbdc8f175","Type":"ContainerDied","Data":"85a045d8bd0d497ded9d38ad1ddf0c4032823afd409fcc7732d2fdbe1e82bac3"} Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.240800 4739 scope.go:117] "RemoveContainer" containerID="38ce4199e7ed95d416cb9e34b0d1992d0babbec85ec2f49c84bf4a889e54ed6d" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.240908 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-l7mfx" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.246346 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vzjc" event={"ID":"cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4","Type":"ContainerDied","Data":"51269b1fdb9ed1e7e6f938e96692530966b85c93102c20b54ab99293fb64ca9c"} Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.246503 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vzjc" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.257267 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" event={"ID":"a480b7ca-0552-40d0-ae91-cccfcba31229","Type":"ContainerStarted","Data":"15bb538be7116a80ecbbdd1f9f5c83dfc49c5d8920bb9bf317740139da67f17b"} Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.257321 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" event={"ID":"a480b7ca-0552-40d0-ae91-cccfcba31229","Type":"ContainerStarted","Data":"ded29a272321643ecfc3c0ed059e43a8d6eb88850e984d757ea7a6a98eed7407"} Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.257823 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.260452 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z5d6l" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.260394 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5d6l" event={"ID":"44224079-e66e-4e6f-a74e-5595d3a378f2","Type":"ContainerDied","Data":"b9ef385aeb24babf07f34492de4a061dfc56e3613a4b034539d5203a3e807332"} Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.260801 4739 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8mjpt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.60:8080/healthz\": dial tcp 10.217.0.60:8080: connect: connection refused" start-of-body= Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.260872 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" podUID="a480b7ca-0552-40d0-ae91-cccfcba31229" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.60:8080/healthz\": dial tcp 10.217.0.60:8080: connect: connection refused" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.268247 4739 scope.go:117] "RemoveContainer" containerID="69ad0151a96195be86ef2057a7e224dabf750cdfcfaf752078ec3f058613a84b" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.273903 4739 generic.go:334] "Generic (PLEG): container finished" podID="83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8" containerID="ff70135236a4e596926d23d394ea708c66ab95f21964efd0cb13000422d59f74" exitCode=0 Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.273963 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ksr6g" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.274017 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ksr6g" event={"ID":"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8","Type":"ContainerDied","Data":"ff70135236a4e596926d23d394ea708c66ab95f21964efd0cb13000422d59f74"} Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.274080 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ksr6g" event={"ID":"83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8","Type":"ContainerDied","Data":"e09b33731de92ceffd80e1b762d140bf5cb8ff1c555c6d7712eed6bfb930e708"} Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.283914 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwqtt" event={"ID":"5b9930dd-3336-438d-84fc-3a62b6a86e75","Type":"ContainerDied","Data":"79f150d4030b41125f63f7869de679675c37692569b6d7e019a14090f7406238"} Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.283999 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwqtt" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.305832 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" podStartSLOduration=2.305804406 podStartE2EDuration="2.305804406s" podCreationTimestamp="2025-12-01 15:40:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:40:08.280114292 +0000 UTC m=+310.105860426" watchObservedRunningTime="2025-12-01 15:40:08.305804406 +0000 UTC m=+310.131550510" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.314155 4739 scope.go:117] "RemoveContainer" containerID="e0707136eda21e51380324d7ffada4ccb73c59b2454ffa2429d3b625410b1f5b" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.321192 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l7mfx"] Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.333379 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l7mfx"] Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.343108 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z5d6l"] Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.347745 4739 scope.go:117] "RemoveContainer" containerID="dad668596b58be1e97b2b883cf9ee74b65331e3802cb846518cf5df2fd724f73" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.365723 4739 scope.go:117] "RemoveContainer" containerID="617ee75a624932ed3767f42d74925f8a0e7f65dd4132310fbdb8d4fdb6607b8f" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.370969 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z5d6l"] Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.378452 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ksr6g"] Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.384773 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ksr6g"] Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.391989 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8vzjc"] Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.414273 4739 scope.go:117] "RemoveContainer" containerID="f991fde905f3ba278301e739122473fa804290580e821bdae59344c3e570444c" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.426513 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8vzjc"] Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.428941 4739 scope.go:117] "RemoveContainer" containerID="47194f78ba84646b6ffdb103252765d9b384c1ffd85546d88d3d6651ce38b357" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.430322 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwqtt"] Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.434314 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nwqtt"] Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.454794 4739 scope.go:117] "RemoveContainer" containerID="ff70135236a4e596926d23d394ea708c66ab95f21964efd0cb13000422d59f74" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.475625 4739 scope.go:117] "RemoveContainer" containerID="7b700d8f868530db267924bb4d5e23c79742cc9babc6ef9b86a223eec950efb4" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.484934 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44224079-e66e-4e6f-a74e-5595d3a378f2" path="/var/lib/kubelet/pods/44224079-e66e-4e6f-a74e-5595d3a378f2/volumes" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.486096 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b9930dd-3336-438d-84fc-3a62b6a86e75" path="/var/lib/kubelet/pods/5b9930dd-3336-438d-84fc-3a62b6a86e75/volumes" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.486652 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8" path="/var/lib/kubelet/pods/83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8/volumes" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.487768 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cbb8f14-1426-4c22-8ccb-f81dbdc8f175" path="/var/lib/kubelet/pods/9cbb8f14-1426-4c22-8ccb-f81dbdc8f175/volumes" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.488370 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" path="/var/lib/kubelet/pods/cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4/volumes" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.494571 4739 scope.go:117] "RemoveContainer" containerID="e4f1ae38209d3755d683eb360ba83057d74322add4086b2c1c6b962a5d8d37c8" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.508552 4739 scope.go:117] "RemoveContainer" containerID="ff70135236a4e596926d23d394ea708c66ab95f21964efd0cb13000422d59f74" Dec 01 15:40:08 crc kubenswrapper[4739]: E1201 15:40:08.509172 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff70135236a4e596926d23d394ea708c66ab95f21964efd0cb13000422d59f74\": container with ID starting with ff70135236a4e596926d23d394ea708c66ab95f21964efd0cb13000422d59f74 not found: ID does not exist" containerID="ff70135236a4e596926d23d394ea708c66ab95f21964efd0cb13000422d59f74" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.509207 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff70135236a4e596926d23d394ea708c66ab95f21964efd0cb13000422d59f74"} err="failed to get container status \"ff70135236a4e596926d23d394ea708c66ab95f21964efd0cb13000422d59f74\": rpc error: code = NotFound desc = could not find container \"ff70135236a4e596926d23d394ea708c66ab95f21964efd0cb13000422d59f74\": container with ID starting with ff70135236a4e596926d23d394ea708c66ab95f21964efd0cb13000422d59f74 not found: ID does not exist" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.509232 4739 scope.go:117] "RemoveContainer" containerID="7b700d8f868530db267924bb4d5e23c79742cc9babc6ef9b86a223eec950efb4" Dec 01 15:40:08 crc kubenswrapper[4739]: E1201 15:40:08.509617 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b700d8f868530db267924bb4d5e23c79742cc9babc6ef9b86a223eec950efb4\": container with ID starting with 7b700d8f868530db267924bb4d5e23c79742cc9babc6ef9b86a223eec950efb4 not found: ID does not exist" containerID="7b700d8f868530db267924bb4d5e23c79742cc9babc6ef9b86a223eec950efb4" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.509645 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b700d8f868530db267924bb4d5e23c79742cc9babc6ef9b86a223eec950efb4"} err="failed to get container status \"7b700d8f868530db267924bb4d5e23c79742cc9babc6ef9b86a223eec950efb4\": rpc error: code = NotFound desc = could not find container \"7b700d8f868530db267924bb4d5e23c79742cc9babc6ef9b86a223eec950efb4\": container with ID starting with 7b700d8f868530db267924bb4d5e23c79742cc9babc6ef9b86a223eec950efb4 not found: ID does not exist" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.509660 4739 scope.go:117] "RemoveContainer" containerID="e4f1ae38209d3755d683eb360ba83057d74322add4086b2c1c6b962a5d8d37c8" Dec 01 15:40:08 crc kubenswrapper[4739]: E1201 15:40:08.511896 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4f1ae38209d3755d683eb360ba83057d74322add4086b2c1c6b962a5d8d37c8\": container with ID starting with e4f1ae38209d3755d683eb360ba83057d74322add4086b2c1c6b962a5d8d37c8 not found: ID does not exist" containerID="e4f1ae38209d3755d683eb360ba83057d74322add4086b2c1c6b962a5d8d37c8" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.511920 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4f1ae38209d3755d683eb360ba83057d74322add4086b2c1c6b962a5d8d37c8"} err="failed to get container status \"e4f1ae38209d3755d683eb360ba83057d74322add4086b2c1c6b962a5d8d37c8\": rpc error: code = NotFound desc = could not find container \"e4f1ae38209d3755d683eb360ba83057d74322add4086b2c1c6b962a5d8d37c8\": container with ID starting with e4f1ae38209d3755d683eb360ba83057d74322add4086b2c1c6b962a5d8d37c8 not found: ID does not exist" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.511936 4739 scope.go:117] "RemoveContainer" containerID="72f42d0ff09f68b5d8669e5dc48d5fef6bae5276fd9dc82aee41e25ce11854e9" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.524130 4739 scope.go:117] "RemoveContainer" containerID="fbd388a660b2093b55c0b4ea8cd1159b870c00492dcee35bc04311bc30aa5f62" Dec 01 15:40:08 crc kubenswrapper[4739]: I1201 15:40:08.536075 4739 scope.go:117] "RemoveContainer" containerID="6797e64177018ca5546315738e5e471866e6af1b2b7dcac7cfa1844e81452155" Dec 01 15:40:09 crc kubenswrapper[4739]: I1201 15:40:09.303499 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-8mjpt" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.441127 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-28lxr"] Dec 01 15:40:17 crc kubenswrapper[4739]: E1201 15:40:17.441881 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9930dd-3336-438d-84fc-3a62b6a86e75" containerName="extract-utilities" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.441893 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9930dd-3336-438d-84fc-3a62b6a86e75" containerName="extract-utilities" Dec 01 15:40:17 crc kubenswrapper[4739]: E1201 15:40:17.441903 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44224079-e66e-4e6f-a74e-5595d3a378f2" containerName="registry-server" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.441909 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="44224079-e66e-4e6f-a74e-5595d3a378f2" containerName="registry-server" Dec 01 15:40:17 crc kubenswrapper[4739]: E1201 15:40:17.441918 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8" containerName="registry-server" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.441924 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8" containerName="registry-server" Dec 01 15:40:17 crc kubenswrapper[4739]: E1201 15:40:17.441936 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cbb8f14-1426-4c22-8ccb-f81dbdc8f175" containerName="marketplace-operator" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.441943 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cbb8f14-1426-4c22-8ccb-f81dbdc8f175" containerName="marketplace-operator" Dec 01 15:40:17 crc kubenswrapper[4739]: E1201 15:40:17.441949 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9930dd-3336-438d-84fc-3a62b6a86e75" containerName="registry-server" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.441955 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9930dd-3336-438d-84fc-3a62b6a86e75" containerName="registry-server" Dec 01 15:40:17 crc kubenswrapper[4739]: E1201 15:40:17.441963 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8" containerName="extract-utilities" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.441968 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8" containerName="extract-utilities" Dec 01 15:40:17 crc kubenswrapper[4739]: E1201 15:40:17.441974 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8" containerName="extract-content" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.441981 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8" containerName="extract-content" Dec 01 15:40:17 crc kubenswrapper[4739]: E1201 15:40:17.441991 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" containerName="extract-utilities" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.442001 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" containerName="extract-utilities" Dec 01 15:40:17 crc kubenswrapper[4739]: E1201 15:40:17.442011 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" containerName="extract-content" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.442018 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" containerName="extract-content" Dec 01 15:40:17 crc kubenswrapper[4739]: E1201 15:40:17.442027 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44224079-e66e-4e6f-a74e-5595d3a378f2" containerName="extract-content" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.442034 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="44224079-e66e-4e6f-a74e-5595d3a378f2" containerName="extract-content" Dec 01 15:40:17 crc kubenswrapper[4739]: E1201 15:40:17.442045 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44224079-e66e-4e6f-a74e-5595d3a378f2" containerName="extract-utilities" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.442051 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="44224079-e66e-4e6f-a74e-5595d3a378f2" containerName="extract-utilities" Dec 01 15:40:17 crc kubenswrapper[4739]: E1201 15:40:17.442060 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cbb8f14-1426-4c22-8ccb-f81dbdc8f175" containerName="marketplace-operator" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.442065 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cbb8f14-1426-4c22-8ccb-f81dbdc8f175" containerName="marketplace-operator" Dec 01 15:40:17 crc kubenswrapper[4739]: E1201 15:40:17.442074 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" containerName="registry-server" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.442079 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" containerName="registry-server" Dec 01 15:40:17 crc kubenswrapper[4739]: E1201 15:40:17.442086 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9930dd-3336-438d-84fc-3a62b6a86e75" containerName="extract-content" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.442093 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9930dd-3336-438d-84fc-3a62b6a86e75" containerName="extract-content" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.442187 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="44224079-e66e-4e6f-a74e-5595d3a378f2" containerName="registry-server" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.442198 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfc6b860-92d2-4c21-b9ec-e7d39b3ad9b4" containerName="registry-server" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.442206 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cbb8f14-1426-4c22-8ccb-f81dbdc8f175" containerName="marketplace-operator" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.442215 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b9930dd-3336-438d-84fc-3a62b6a86e75" containerName="registry-server" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.442223 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cbb8f14-1426-4c22-8ccb-f81dbdc8f175" containerName="marketplace-operator" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.442232 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="83a6fb61-1d84-43a3-bc8e-5890c3c0a4f8" containerName="registry-server" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.442899 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-28lxr" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.445739 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.474192 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-28lxr"] Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.520390 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40475f1-4984-4d14-95ff-9ee7f6493756-utilities\") pod \"redhat-marketplace-28lxr\" (UID: \"a40475f1-4984-4d14-95ff-9ee7f6493756\") " pod="openshift-marketplace/redhat-marketplace-28lxr" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.520690 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40475f1-4984-4d14-95ff-9ee7f6493756-catalog-content\") pod \"redhat-marketplace-28lxr\" (UID: \"a40475f1-4984-4d14-95ff-9ee7f6493756\") " pod="openshift-marketplace/redhat-marketplace-28lxr" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.520786 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txwnz\" (UniqueName: \"kubernetes.io/projected/a40475f1-4984-4d14-95ff-9ee7f6493756-kube-api-access-txwnz\") pod \"redhat-marketplace-28lxr\" (UID: \"a40475f1-4984-4d14-95ff-9ee7f6493756\") " pod="openshift-marketplace/redhat-marketplace-28lxr" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.622437 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40475f1-4984-4d14-95ff-9ee7f6493756-catalog-content\") pod \"redhat-marketplace-28lxr\" (UID: \"a40475f1-4984-4d14-95ff-9ee7f6493756\") " pod="openshift-marketplace/redhat-marketplace-28lxr" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.622485 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txwnz\" (UniqueName: \"kubernetes.io/projected/a40475f1-4984-4d14-95ff-9ee7f6493756-kube-api-access-txwnz\") pod \"redhat-marketplace-28lxr\" (UID: \"a40475f1-4984-4d14-95ff-9ee7f6493756\") " pod="openshift-marketplace/redhat-marketplace-28lxr" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.622546 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40475f1-4984-4d14-95ff-9ee7f6493756-utilities\") pod \"redhat-marketplace-28lxr\" (UID: \"a40475f1-4984-4d14-95ff-9ee7f6493756\") " pod="openshift-marketplace/redhat-marketplace-28lxr" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.622997 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40475f1-4984-4d14-95ff-9ee7f6493756-utilities\") pod \"redhat-marketplace-28lxr\" (UID: \"a40475f1-4984-4d14-95ff-9ee7f6493756\") " pod="openshift-marketplace/redhat-marketplace-28lxr" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.623209 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40475f1-4984-4d14-95ff-9ee7f6493756-catalog-content\") pod \"redhat-marketplace-28lxr\" (UID: \"a40475f1-4984-4d14-95ff-9ee7f6493756\") " pod="openshift-marketplace/redhat-marketplace-28lxr" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.641125 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fnwbc"] Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.644111 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.648748 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.657565 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fnwbc"] Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.659195 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txwnz\" (UniqueName: \"kubernetes.io/projected/a40475f1-4984-4d14-95ff-9ee7f6493756-kube-api-access-txwnz\") pod \"redhat-marketplace-28lxr\" (UID: \"a40475f1-4984-4d14-95ff-9ee7f6493756\") " pod="openshift-marketplace/redhat-marketplace-28lxr" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.723458 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgqw5\" (UniqueName: \"kubernetes.io/projected/434fbbdb-10c8-461e-812f-7725b27b0111-kube-api-access-kgqw5\") pod \"redhat-operators-fnwbc\" (UID: \"434fbbdb-10c8-461e-812f-7725b27b0111\") " pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.723547 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434fbbdb-10c8-461e-812f-7725b27b0111-catalog-content\") pod \"redhat-operators-fnwbc\" (UID: \"434fbbdb-10c8-461e-812f-7725b27b0111\") " pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.723591 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434fbbdb-10c8-461e-812f-7725b27b0111-utilities\") pod \"redhat-operators-fnwbc\" (UID: \"434fbbdb-10c8-461e-812f-7725b27b0111\") " pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.768457 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-28lxr" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.831171 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgqw5\" (UniqueName: \"kubernetes.io/projected/434fbbdb-10c8-461e-812f-7725b27b0111-kube-api-access-kgqw5\") pod \"redhat-operators-fnwbc\" (UID: \"434fbbdb-10c8-461e-812f-7725b27b0111\") " pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.831260 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434fbbdb-10c8-461e-812f-7725b27b0111-catalog-content\") pod \"redhat-operators-fnwbc\" (UID: \"434fbbdb-10c8-461e-812f-7725b27b0111\") " pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.831307 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434fbbdb-10c8-461e-812f-7725b27b0111-utilities\") pod \"redhat-operators-fnwbc\" (UID: \"434fbbdb-10c8-461e-812f-7725b27b0111\") " pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.832286 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434fbbdb-10c8-461e-812f-7725b27b0111-utilities\") pod \"redhat-operators-fnwbc\" (UID: \"434fbbdb-10c8-461e-812f-7725b27b0111\") " pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.832446 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434fbbdb-10c8-461e-812f-7725b27b0111-catalog-content\") pod \"redhat-operators-fnwbc\" (UID: \"434fbbdb-10c8-461e-812f-7725b27b0111\") " pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.853626 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgqw5\" (UniqueName: \"kubernetes.io/projected/434fbbdb-10c8-461e-812f-7725b27b0111-kube-api-access-kgqw5\") pod \"redhat-operators-fnwbc\" (UID: \"434fbbdb-10c8-461e-812f-7725b27b0111\") " pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:40:17 crc kubenswrapper[4739]: I1201 15:40:17.987705 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:40:18 crc kubenswrapper[4739]: I1201 15:40:18.189922 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-28lxr"] Dec 01 15:40:18 crc kubenswrapper[4739]: I1201 15:40:18.360946 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28lxr" event={"ID":"a40475f1-4984-4d14-95ff-9ee7f6493756","Type":"ContainerStarted","Data":"4d7fd347357143e91ca03454d2ce76ea621b2a3f5fb9abec43b011f286f6e82a"} Dec 01 15:40:18 crc kubenswrapper[4739]: I1201 15:40:18.360985 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28lxr" event={"ID":"a40475f1-4984-4d14-95ff-9ee7f6493756","Type":"ContainerStarted","Data":"35a318af73af7be161a67935f53d0cfc8acde8ccb02c170daa17af2d11adbb28"} Dec 01 15:40:18 crc kubenswrapper[4739]: I1201 15:40:18.381311 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fnwbc"] Dec 01 15:40:18 crc kubenswrapper[4739]: W1201 15:40:18.412966 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod434fbbdb_10c8_461e_812f_7725b27b0111.slice/crio-a1ca7b89dddc1a98f6adc160d57374f4628af2f72b0cf7a3853f6b2e7dd4d217 WatchSource:0}: Error finding container a1ca7b89dddc1a98f6adc160d57374f4628af2f72b0cf7a3853f6b2e7dd4d217: Status 404 returned error can't find the container with id a1ca7b89dddc1a98f6adc160d57374f4628af2f72b0cf7a3853f6b2e7dd4d217 Dec 01 15:40:19 crc kubenswrapper[4739]: I1201 15:40:19.368166 4739 generic.go:334] "Generic (PLEG): container finished" podID="434fbbdb-10c8-461e-812f-7725b27b0111" containerID="60b7ed89dcd28ad8ff68122ff3a20ac1f5bebaaf5345959c8d57aab735b16d6f" exitCode=0 Dec 01 15:40:19 crc kubenswrapper[4739]: I1201 15:40:19.368229 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fnwbc" event={"ID":"434fbbdb-10c8-461e-812f-7725b27b0111","Type":"ContainerDied","Data":"60b7ed89dcd28ad8ff68122ff3a20ac1f5bebaaf5345959c8d57aab735b16d6f"} Dec 01 15:40:19 crc kubenswrapper[4739]: I1201 15:40:19.368675 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fnwbc" event={"ID":"434fbbdb-10c8-461e-812f-7725b27b0111","Type":"ContainerStarted","Data":"a1ca7b89dddc1a98f6adc160d57374f4628af2f72b0cf7a3853f6b2e7dd4d217"} Dec 01 15:40:19 crc kubenswrapper[4739]: I1201 15:40:19.370015 4739 generic.go:334] "Generic (PLEG): container finished" podID="a40475f1-4984-4d14-95ff-9ee7f6493756" containerID="4d7fd347357143e91ca03454d2ce76ea621b2a3f5fb9abec43b011f286f6e82a" exitCode=0 Dec 01 15:40:19 crc kubenswrapper[4739]: I1201 15:40:19.370041 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28lxr" event={"ID":"a40475f1-4984-4d14-95ff-9ee7f6493756","Type":"ContainerDied","Data":"4d7fd347357143e91ca03454d2ce76ea621b2a3f5fb9abec43b011f286f6e82a"} Dec 01 15:40:19 crc kubenswrapper[4739]: I1201 15:40:19.866399 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9jfkx"] Dec 01 15:40:19 crc kubenswrapper[4739]: I1201 15:40:19.875769 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9jfkx"] Dec 01 15:40:19 crc kubenswrapper[4739]: I1201 15:40:19.875874 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 15:40:19 crc kubenswrapper[4739]: I1201 15:40:19.878384 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 15:40:19 crc kubenswrapper[4739]: I1201 15:40:19.958233 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqvnj\" (UniqueName: \"kubernetes.io/projected/c75bb144-38de-46c1-af46-26a0840c1632-kube-api-access-zqvnj\") pod \"certified-operators-9jfkx\" (UID: \"c75bb144-38de-46c1-af46-26a0840c1632\") " pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 15:40:19 crc kubenswrapper[4739]: I1201 15:40:19.958276 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c75bb144-38de-46c1-af46-26a0840c1632-catalog-content\") pod \"certified-operators-9jfkx\" (UID: \"c75bb144-38de-46c1-af46-26a0840c1632\") " pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 15:40:19 crc kubenswrapper[4739]: I1201 15:40:19.958309 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c75bb144-38de-46c1-af46-26a0840c1632-utilities\") pod \"certified-operators-9jfkx\" (UID: \"c75bb144-38de-46c1-af46-26a0840c1632\") " pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.038260 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-52jpf"] Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.039175 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-52jpf" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.040503 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.052661 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-52jpf"] Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.058747 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqvnj\" (UniqueName: \"kubernetes.io/projected/c75bb144-38de-46c1-af46-26a0840c1632-kube-api-access-zqvnj\") pod \"certified-operators-9jfkx\" (UID: \"c75bb144-38de-46c1-af46-26a0840c1632\") " pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.058799 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c75bb144-38de-46c1-af46-26a0840c1632-catalog-content\") pod \"certified-operators-9jfkx\" (UID: \"c75bb144-38de-46c1-af46-26a0840c1632\") " pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.058829 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7tvw\" (UniqueName: \"kubernetes.io/projected/079aa9ef-3896-4f11-9a9f-e8cb26607b88-kube-api-access-v7tvw\") pod \"community-operators-52jpf\" (UID: \"079aa9ef-3896-4f11-9a9f-e8cb26607b88\") " pod="openshift-marketplace/community-operators-52jpf" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.058867 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c75bb144-38de-46c1-af46-26a0840c1632-utilities\") pod \"certified-operators-9jfkx\" (UID: \"c75bb144-38de-46c1-af46-26a0840c1632\") " pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.058901 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/079aa9ef-3896-4f11-9a9f-e8cb26607b88-catalog-content\") pod \"community-operators-52jpf\" (UID: \"079aa9ef-3896-4f11-9a9f-e8cb26607b88\") " pod="openshift-marketplace/community-operators-52jpf" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.058948 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/079aa9ef-3896-4f11-9a9f-e8cb26607b88-utilities\") pod \"community-operators-52jpf\" (UID: \"079aa9ef-3896-4f11-9a9f-e8cb26607b88\") " pod="openshift-marketplace/community-operators-52jpf" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.059780 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c75bb144-38de-46c1-af46-26a0840c1632-catalog-content\") pod \"certified-operators-9jfkx\" (UID: \"c75bb144-38de-46c1-af46-26a0840c1632\") " pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.060056 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c75bb144-38de-46c1-af46-26a0840c1632-utilities\") pod \"certified-operators-9jfkx\" (UID: \"c75bb144-38de-46c1-af46-26a0840c1632\") " pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.085122 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqvnj\" (UniqueName: \"kubernetes.io/projected/c75bb144-38de-46c1-af46-26a0840c1632-kube-api-access-zqvnj\") pod \"certified-operators-9jfkx\" (UID: \"c75bb144-38de-46c1-af46-26a0840c1632\") " pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.159612 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/079aa9ef-3896-4f11-9a9f-e8cb26607b88-utilities\") pod \"community-operators-52jpf\" (UID: \"079aa9ef-3896-4f11-9a9f-e8cb26607b88\") " pod="openshift-marketplace/community-operators-52jpf" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.159689 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7tvw\" (UniqueName: \"kubernetes.io/projected/079aa9ef-3896-4f11-9a9f-e8cb26607b88-kube-api-access-v7tvw\") pod \"community-operators-52jpf\" (UID: \"079aa9ef-3896-4f11-9a9f-e8cb26607b88\") " pod="openshift-marketplace/community-operators-52jpf" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.159723 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/079aa9ef-3896-4f11-9a9f-e8cb26607b88-catalog-content\") pod \"community-operators-52jpf\" (UID: \"079aa9ef-3896-4f11-9a9f-e8cb26607b88\") " pod="openshift-marketplace/community-operators-52jpf" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.160113 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/079aa9ef-3896-4f11-9a9f-e8cb26607b88-utilities\") pod \"community-operators-52jpf\" (UID: \"079aa9ef-3896-4f11-9a9f-e8cb26607b88\") " pod="openshift-marketplace/community-operators-52jpf" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.160134 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/079aa9ef-3896-4f11-9a9f-e8cb26607b88-catalog-content\") pod \"community-operators-52jpf\" (UID: \"079aa9ef-3896-4f11-9a9f-e8cb26607b88\") " pod="openshift-marketplace/community-operators-52jpf" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.200679 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7tvw\" (UniqueName: \"kubernetes.io/projected/079aa9ef-3896-4f11-9a9f-e8cb26607b88-kube-api-access-v7tvw\") pod \"community-operators-52jpf\" (UID: \"079aa9ef-3896-4f11-9a9f-e8cb26607b88\") " pod="openshift-marketplace/community-operators-52jpf" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.201952 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.367748 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-52jpf" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.376762 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28lxr" event={"ID":"a40475f1-4984-4d14-95ff-9ee7f6493756","Type":"ContainerStarted","Data":"db5e9fa56cc7b63810eb22b44737e51cce2a2d5355c53418f9bb4a802d99195d"} Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.594259 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9jfkx"] Dec 01 15:40:20 crc kubenswrapper[4739]: W1201 15:40:20.600493 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc75bb144_38de_46c1_af46_26a0840c1632.slice/crio-f61198468e66eb5397719799d2e765b393abfb227f84514cee345be3210b0c07 WatchSource:0}: Error finding container f61198468e66eb5397719799d2e765b393abfb227f84514cee345be3210b0c07: Status 404 returned error can't find the container with id f61198468e66eb5397719799d2e765b393abfb227f84514cee345be3210b0c07 Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.769807 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-52jpf"] Dec 01 15:40:20 crc kubenswrapper[4739]: W1201 15:40:20.785377 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod079aa9ef_3896_4f11_9a9f_e8cb26607b88.slice/crio-c7d89dc3eb3eaa062e4982aa42686cc4051cd923af4fccac4119629a6caff56d WatchSource:0}: Error finding container c7d89dc3eb3eaa062e4982aa42686cc4051cd923af4fccac4119629a6caff56d: Status 404 returned error can't find the container with id c7d89dc3eb3eaa062e4982aa42686cc4051cd923af4fccac4119629a6caff56d Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.800350 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bzzg4"] Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.801168 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.824927 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bzzg4"] Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.974490 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.974591 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7049190e-f905-470d-a9f4-019dc008f763-registry-certificates\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.974646 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7049190e-f905-470d-a9f4-019dc008f763-trusted-ca\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.974678 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7049190e-f905-470d-a9f4-019dc008f763-bound-sa-token\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.974772 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7049190e-f905-470d-a9f4-019dc008f763-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.974815 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlh2z\" (UniqueName: \"kubernetes.io/projected/7049190e-f905-470d-a9f4-019dc008f763-kube-api-access-zlh2z\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.974864 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7049190e-f905-470d-a9f4-019dc008f763-registry-tls\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:20 crc kubenswrapper[4739]: I1201 15:40:20.974960 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7049190e-f905-470d-a9f4-019dc008f763-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:20.999994 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.076357 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7049190e-f905-470d-a9f4-019dc008f763-registry-certificates\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.076434 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7049190e-f905-470d-a9f4-019dc008f763-trusted-ca\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.076463 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7049190e-f905-470d-a9f4-019dc008f763-bound-sa-token\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.076486 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7049190e-f905-470d-a9f4-019dc008f763-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.076505 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlh2z\" (UniqueName: \"kubernetes.io/projected/7049190e-f905-470d-a9f4-019dc008f763-kube-api-access-zlh2z\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.076520 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7049190e-f905-470d-a9f4-019dc008f763-registry-tls\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.076550 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7049190e-f905-470d-a9f4-019dc008f763-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.078226 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7049190e-f905-470d-a9f4-019dc008f763-trusted-ca\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.078313 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7049190e-f905-470d-a9f4-019dc008f763-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.078554 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7049190e-f905-470d-a9f4-019dc008f763-registry-certificates\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.083114 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7049190e-f905-470d-a9f4-019dc008f763-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.083604 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7049190e-f905-470d-a9f4-019dc008f763-registry-tls\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.094063 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlh2z\" (UniqueName: \"kubernetes.io/projected/7049190e-f905-470d-a9f4-019dc008f763-kube-api-access-zlh2z\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.100615 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7049190e-f905-470d-a9f4-019dc008f763-bound-sa-token\") pod \"image-registry-66df7c8f76-bzzg4\" (UID: \"7049190e-f905-470d-a9f4-019dc008f763\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.130836 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.392575 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fnwbc" event={"ID":"434fbbdb-10c8-461e-812f-7725b27b0111","Type":"ContainerStarted","Data":"d707e6fd0862a55ef17b0c0dda043f2301af28672b80c8c03e1fd2921f9114b3"} Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.393753 4739 generic.go:334] "Generic (PLEG): container finished" podID="c75bb144-38de-46c1-af46-26a0840c1632" containerID="69e528e161294d1152ab13b170a3e95cb478bf3d043d7ec43068059a3cbfddf4" exitCode=0 Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.393807 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jfkx" event={"ID":"c75bb144-38de-46c1-af46-26a0840c1632","Type":"ContainerDied","Data":"69e528e161294d1152ab13b170a3e95cb478bf3d043d7ec43068059a3cbfddf4"} Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.393832 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jfkx" event={"ID":"c75bb144-38de-46c1-af46-26a0840c1632","Type":"ContainerStarted","Data":"f61198468e66eb5397719799d2e765b393abfb227f84514cee345be3210b0c07"} Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.395937 4739 generic.go:334] "Generic (PLEG): container finished" podID="079aa9ef-3896-4f11-9a9f-e8cb26607b88" containerID="4391f9750ac51234dc0a16cfb7b8e2c1938e004783714af7ec9bc68170f87edd" exitCode=0 Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.396013 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52jpf" event={"ID":"079aa9ef-3896-4f11-9a9f-e8cb26607b88","Type":"ContainerDied","Data":"4391f9750ac51234dc0a16cfb7b8e2c1938e004783714af7ec9bc68170f87edd"} Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.396045 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52jpf" event={"ID":"079aa9ef-3896-4f11-9a9f-e8cb26607b88","Type":"ContainerStarted","Data":"c7d89dc3eb3eaa062e4982aa42686cc4051cd923af4fccac4119629a6caff56d"} Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.402408 4739 generic.go:334] "Generic (PLEG): container finished" podID="a40475f1-4984-4d14-95ff-9ee7f6493756" containerID="db5e9fa56cc7b63810eb22b44737e51cce2a2d5355c53418f9bb4a802d99195d" exitCode=0 Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.402489 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28lxr" event={"ID":"a40475f1-4984-4d14-95ff-9ee7f6493756","Type":"ContainerDied","Data":"db5e9fa56cc7b63810eb22b44737e51cce2a2d5355c53418f9bb4a802d99195d"} Dec 01 15:40:21 crc kubenswrapper[4739]: I1201 15:40:21.545745 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bzzg4"] Dec 01 15:40:22 crc kubenswrapper[4739]: I1201 15:40:22.408015 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" event={"ID":"7049190e-f905-470d-a9f4-019dc008f763","Type":"ContainerStarted","Data":"505d45ef3714d7831b63211984991c906e10c5c301e1575a8fc971adb7411707"} Dec 01 15:40:22 crc kubenswrapper[4739]: I1201 15:40:22.408805 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" event={"ID":"7049190e-f905-470d-a9f4-019dc008f763","Type":"ContainerStarted","Data":"7dc50783e73126fd522852e51584163e00ef3da341f2f61dcd5660dd78fd6597"} Dec 01 15:40:22 crc kubenswrapper[4739]: I1201 15:40:22.409112 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:22 crc kubenswrapper[4739]: I1201 15:40:22.411396 4739 generic.go:334] "Generic (PLEG): container finished" podID="079aa9ef-3896-4f11-9a9f-e8cb26607b88" containerID="efd49d28edb13c19a93d0059d68d8e724e4724e47798fe264251652b4edd22f6" exitCode=0 Dec 01 15:40:22 crc kubenswrapper[4739]: I1201 15:40:22.411776 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52jpf" event={"ID":"079aa9ef-3896-4f11-9a9f-e8cb26607b88","Type":"ContainerDied","Data":"efd49d28edb13c19a93d0059d68d8e724e4724e47798fe264251652b4edd22f6"} Dec 01 15:40:22 crc kubenswrapper[4739]: I1201 15:40:22.414263 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28lxr" event={"ID":"a40475f1-4984-4d14-95ff-9ee7f6493756","Type":"ContainerStarted","Data":"9cfb72347715aa0a2d372f6addb28d13b2101efc75a9ce9ace15d92165f306b9"} Dec 01 15:40:22 crc kubenswrapper[4739]: I1201 15:40:22.417294 4739 generic.go:334] "Generic (PLEG): container finished" podID="434fbbdb-10c8-461e-812f-7725b27b0111" containerID="d707e6fd0862a55ef17b0c0dda043f2301af28672b80c8c03e1fd2921f9114b3" exitCode=0 Dec 01 15:40:22 crc kubenswrapper[4739]: I1201 15:40:22.417891 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fnwbc" event={"ID":"434fbbdb-10c8-461e-812f-7725b27b0111","Type":"ContainerDied","Data":"d707e6fd0862a55ef17b0c0dda043f2301af28672b80c8c03e1fd2921f9114b3"} Dec 01 15:40:22 crc kubenswrapper[4739]: I1201 15:40:22.434187 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" podStartSLOduration=2.434172498 podStartE2EDuration="2.434172498s" podCreationTimestamp="2025-12-01 15:40:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:40:22.43298113 +0000 UTC m=+324.258727244" watchObservedRunningTime="2025-12-01 15:40:22.434172498 +0000 UTC m=+324.259918592" Dec 01 15:40:22 crc kubenswrapper[4739]: I1201 15:40:22.476672 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-28lxr" podStartSLOduration=3.02915513 podStartE2EDuration="5.47665528s" podCreationTimestamp="2025-12-01 15:40:17 +0000 UTC" firstStartedPulling="2025-12-01 15:40:19.37149101 +0000 UTC m=+321.197237144" lastFinishedPulling="2025-12-01 15:40:21.8189912 +0000 UTC m=+323.644737294" observedRunningTime="2025-12-01 15:40:22.475357028 +0000 UTC m=+324.301103122" watchObservedRunningTime="2025-12-01 15:40:22.47665528 +0000 UTC m=+324.302401374" Dec 01 15:40:23 crc kubenswrapper[4739]: I1201 15:40:23.424568 4739 generic.go:334] "Generic (PLEG): container finished" podID="c75bb144-38de-46c1-af46-26a0840c1632" containerID="b9fd5198001b68bbdd90bf678024d7611af26b45788e6ad924f74c234412bfc4" exitCode=0 Dec 01 15:40:23 crc kubenswrapper[4739]: I1201 15:40:23.424662 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jfkx" event={"ID":"c75bb144-38de-46c1-af46-26a0840c1632","Type":"ContainerDied","Data":"b9fd5198001b68bbdd90bf678024d7611af26b45788e6ad924f74c234412bfc4"} Dec 01 15:40:23 crc kubenswrapper[4739]: I1201 15:40:23.429218 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52jpf" event={"ID":"079aa9ef-3896-4f11-9a9f-e8cb26607b88","Type":"ContainerStarted","Data":"77db31bd9f4b2ea11b8cb617f98352ce305dbc6cf270e1306446874327e7d00d"} Dec 01 15:40:23 crc kubenswrapper[4739]: I1201 15:40:23.432800 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fnwbc" event={"ID":"434fbbdb-10c8-461e-812f-7725b27b0111","Type":"ContainerStarted","Data":"dedcf54eafbcff7957421b1569934a5e4a9c7137a06b70c8c026ebfd4e63a614"} Dec 01 15:40:23 crc kubenswrapper[4739]: I1201 15:40:23.466276 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-52jpf" podStartSLOduration=1.9043446670000002 podStartE2EDuration="3.46625823s" podCreationTimestamp="2025-12-01 15:40:20 +0000 UTC" firstStartedPulling="2025-12-01 15:40:21.39718083 +0000 UTC m=+323.222926954" lastFinishedPulling="2025-12-01 15:40:22.959094423 +0000 UTC m=+324.784840517" observedRunningTime="2025-12-01 15:40:23.462190279 +0000 UTC m=+325.287936373" watchObservedRunningTime="2025-12-01 15:40:23.46625823 +0000 UTC m=+325.292004324" Dec 01 15:40:23 crc kubenswrapper[4739]: I1201 15:40:23.490214 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fnwbc" podStartSLOduration=2.956851974 podStartE2EDuration="6.490184756s" podCreationTimestamp="2025-12-01 15:40:17 +0000 UTC" firstStartedPulling="2025-12-01 15:40:19.370488509 +0000 UTC m=+321.196234643" lastFinishedPulling="2025-12-01 15:40:22.903821311 +0000 UTC m=+324.729567425" observedRunningTime="2025-12-01 15:40:23.485809096 +0000 UTC m=+325.311555190" watchObservedRunningTime="2025-12-01 15:40:23.490184756 +0000 UTC m=+325.315930890" Dec 01 15:40:24 crc kubenswrapper[4739]: I1201 15:40:24.440127 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jfkx" event={"ID":"c75bb144-38de-46c1-af46-26a0840c1632","Type":"ContainerStarted","Data":"5e04d551c5e71170b416dcb611a70084a5a0d915e93c76384f774a86d0f2250b"} Dec 01 15:40:24 crc kubenswrapper[4739]: I1201 15:40:24.463020 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9jfkx" podStartSLOduration=2.991164348 podStartE2EDuration="5.463004637s" podCreationTimestamp="2025-12-01 15:40:19 +0000 UTC" firstStartedPulling="2025-12-01 15:40:21.401380234 +0000 UTC m=+323.227126368" lastFinishedPulling="2025-12-01 15:40:23.873220553 +0000 UTC m=+325.698966657" observedRunningTime="2025-12-01 15:40:24.461654784 +0000 UTC m=+326.287400878" watchObservedRunningTime="2025-12-01 15:40:24.463004637 +0000 UTC m=+326.288750731" Dec 01 15:40:27 crc kubenswrapper[4739]: I1201 15:40:27.730933 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-bf64b4967-gbbbn"] Dec 01 15:40:27 crc kubenswrapper[4739]: I1201 15:40:27.731472 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" podUID="20f7f236-e00a-4002-9056-5099d2a6e485" containerName="controller-manager" containerID="cri-o://3cb290daeb34ddfc00da0252d0c45471ff170aa7e2971968d423e25871284cdb" gracePeriod=30 Dec 01 15:40:27 crc kubenswrapper[4739]: I1201 15:40:27.769144 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-28lxr" Dec 01 15:40:27 crc kubenswrapper[4739]: I1201 15:40:27.769543 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-28lxr" Dec 01 15:40:27 crc kubenswrapper[4739]: I1201 15:40:27.821442 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-28lxr" Dec 01 15:40:27 crc kubenswrapper[4739]: I1201 15:40:27.988470 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:40:27 crc kubenswrapper[4739]: I1201 15:40:27.988880 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.461121 4739 generic.go:334] "Generic (PLEG): container finished" podID="20f7f236-e00a-4002-9056-5099d2a6e485" containerID="3cb290daeb34ddfc00da0252d0c45471ff170aa7e2971968d423e25871284cdb" exitCode=0 Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.461214 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" event={"ID":"20f7f236-e00a-4002-9056-5099d2a6e485","Type":"ContainerDied","Data":"3cb290daeb34ddfc00da0252d0c45471ff170aa7e2971968d423e25871284cdb"} Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.505600 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-28lxr" Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.667596 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.785442 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20f7f236-e00a-4002-9056-5099d2a6e485-serving-cert\") pod \"20f7f236-e00a-4002-9056-5099d2a6e485\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.785530 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9cch\" (UniqueName: \"kubernetes.io/projected/20f7f236-e00a-4002-9056-5099d2a6e485-kube-api-access-z9cch\") pod \"20f7f236-e00a-4002-9056-5099d2a6e485\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.785563 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-proxy-ca-bundles\") pod \"20f7f236-e00a-4002-9056-5099d2a6e485\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.785597 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-config\") pod \"20f7f236-e00a-4002-9056-5099d2a6e485\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.785638 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-client-ca\") pod \"20f7f236-e00a-4002-9056-5099d2a6e485\" (UID: \"20f7f236-e00a-4002-9056-5099d2a6e485\") " Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.786804 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-client-ca" (OuterVolumeSpecName: "client-ca") pod "20f7f236-e00a-4002-9056-5099d2a6e485" (UID: "20f7f236-e00a-4002-9056-5099d2a6e485"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.787234 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "20f7f236-e00a-4002-9056-5099d2a6e485" (UID: "20f7f236-e00a-4002-9056-5099d2a6e485"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.787531 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-config" (OuterVolumeSpecName: "config") pod "20f7f236-e00a-4002-9056-5099d2a6e485" (UID: "20f7f236-e00a-4002-9056-5099d2a6e485"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.791473 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20f7f236-e00a-4002-9056-5099d2a6e485-kube-api-access-z9cch" (OuterVolumeSpecName: "kube-api-access-z9cch") pod "20f7f236-e00a-4002-9056-5099d2a6e485" (UID: "20f7f236-e00a-4002-9056-5099d2a6e485"). InnerVolumeSpecName "kube-api-access-z9cch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.793168 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20f7f236-e00a-4002-9056-5099d2a6e485-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "20f7f236-e00a-4002-9056-5099d2a6e485" (UID: "20f7f236-e00a-4002-9056-5099d2a6e485"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.887165 4739 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.887200 4739 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20f7f236-e00a-4002-9056-5099d2a6e485-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.887211 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9cch\" (UniqueName: \"kubernetes.io/projected/20f7f236-e00a-4002-9056-5099d2a6e485-kube-api-access-z9cch\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.887221 4739 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:28 crc kubenswrapper[4739]: I1201 15:40:28.887230 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20f7f236-e00a-4002-9056-5099d2a6e485-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.029749 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fnwbc" podUID="434fbbdb-10c8-461e-812f-7725b27b0111" containerName="registry-server" probeResult="failure" output=< Dec 01 15:40:29 crc kubenswrapper[4739]: timeout: failed to connect service ":50051" within 1s Dec 01 15:40:29 crc kubenswrapper[4739]: > Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.297982 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5598bcd6b-dsbdm"] Dec 01 15:40:29 crc kubenswrapper[4739]: E1201 15:40:29.298176 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20f7f236-e00a-4002-9056-5099d2a6e485" containerName="controller-manager" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.298187 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="20f7f236-e00a-4002-9056-5099d2a6e485" containerName="controller-manager" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.298273 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="20f7f236-e00a-4002-9056-5099d2a6e485" containerName="controller-manager" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.298629 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.314680 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5598bcd6b-dsbdm"] Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.469968 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.470034 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-bf64b4967-gbbbn" event={"ID":"20f7f236-e00a-4002-9056-5099d2a6e485","Type":"ContainerDied","Data":"b486cb9f887625bc111ae3cf1e3ff7664aa89c6706420a86493da17c43553c3e"} Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.470081 4739 scope.go:117] "RemoveContainer" containerID="3cb290daeb34ddfc00da0252d0c45471ff170aa7e2971968d423e25871284cdb" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.494325 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5433eac9-61a8-47e5-ba62-c9e09bcfe50b-client-ca\") pod \"controller-manager-5598bcd6b-dsbdm\" (UID: \"5433eac9-61a8-47e5-ba62-c9e09bcfe50b\") " pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.494635 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4qjz\" (UniqueName: \"kubernetes.io/projected/5433eac9-61a8-47e5-ba62-c9e09bcfe50b-kube-api-access-g4qjz\") pod \"controller-manager-5598bcd6b-dsbdm\" (UID: \"5433eac9-61a8-47e5-ba62-c9e09bcfe50b\") " pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.494732 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5433eac9-61a8-47e5-ba62-c9e09bcfe50b-proxy-ca-bundles\") pod \"controller-manager-5598bcd6b-dsbdm\" (UID: \"5433eac9-61a8-47e5-ba62-c9e09bcfe50b\") " pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.494837 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5433eac9-61a8-47e5-ba62-c9e09bcfe50b-serving-cert\") pod \"controller-manager-5598bcd6b-dsbdm\" (UID: \"5433eac9-61a8-47e5-ba62-c9e09bcfe50b\") " pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.495272 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5433eac9-61a8-47e5-ba62-c9e09bcfe50b-config\") pod \"controller-manager-5598bcd6b-dsbdm\" (UID: \"5433eac9-61a8-47e5-ba62-c9e09bcfe50b\") " pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.509025 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-bf64b4967-gbbbn"] Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.514970 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-bf64b4967-gbbbn"] Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.597092 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4qjz\" (UniqueName: \"kubernetes.io/projected/5433eac9-61a8-47e5-ba62-c9e09bcfe50b-kube-api-access-g4qjz\") pod \"controller-manager-5598bcd6b-dsbdm\" (UID: \"5433eac9-61a8-47e5-ba62-c9e09bcfe50b\") " pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.598360 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5433eac9-61a8-47e5-ba62-c9e09bcfe50b-proxy-ca-bundles\") pod \"controller-manager-5598bcd6b-dsbdm\" (UID: \"5433eac9-61a8-47e5-ba62-c9e09bcfe50b\") " pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.598691 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5433eac9-61a8-47e5-ba62-c9e09bcfe50b-serving-cert\") pod \"controller-manager-5598bcd6b-dsbdm\" (UID: \"5433eac9-61a8-47e5-ba62-c9e09bcfe50b\") " pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.599007 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5433eac9-61a8-47e5-ba62-c9e09bcfe50b-config\") pod \"controller-manager-5598bcd6b-dsbdm\" (UID: \"5433eac9-61a8-47e5-ba62-c9e09bcfe50b\") " pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.599404 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5433eac9-61a8-47e5-ba62-c9e09bcfe50b-client-ca\") pod \"controller-manager-5598bcd6b-dsbdm\" (UID: \"5433eac9-61a8-47e5-ba62-c9e09bcfe50b\") " pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.600895 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5433eac9-61a8-47e5-ba62-c9e09bcfe50b-client-ca\") pod \"controller-manager-5598bcd6b-dsbdm\" (UID: \"5433eac9-61a8-47e5-ba62-c9e09bcfe50b\") " pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.600950 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5433eac9-61a8-47e5-ba62-c9e09bcfe50b-proxy-ca-bundles\") pod \"controller-manager-5598bcd6b-dsbdm\" (UID: \"5433eac9-61a8-47e5-ba62-c9e09bcfe50b\") " pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.606625 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5433eac9-61a8-47e5-ba62-c9e09bcfe50b-config\") pod \"controller-manager-5598bcd6b-dsbdm\" (UID: \"5433eac9-61a8-47e5-ba62-c9e09bcfe50b\") " pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.615328 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5433eac9-61a8-47e5-ba62-c9e09bcfe50b-serving-cert\") pod \"controller-manager-5598bcd6b-dsbdm\" (UID: \"5433eac9-61a8-47e5-ba62-c9e09bcfe50b\") " pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.625227 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4qjz\" (UniqueName: \"kubernetes.io/projected/5433eac9-61a8-47e5-ba62-c9e09bcfe50b-kube-api-access-g4qjz\") pod \"controller-manager-5598bcd6b-dsbdm\" (UID: \"5433eac9-61a8-47e5-ba62-c9e09bcfe50b\") " pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:29 crc kubenswrapper[4739]: I1201 15:40:29.915014 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:30 crc kubenswrapper[4739]: I1201 15:40:30.201387 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5598bcd6b-dsbdm"] Dec 01 15:40:30 crc kubenswrapper[4739]: I1201 15:40:30.202736 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 15:40:30 crc kubenswrapper[4739]: I1201 15:40:30.202834 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 15:40:30 crc kubenswrapper[4739]: I1201 15:40:30.246760 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 15:40:30 crc kubenswrapper[4739]: I1201 15:40:30.368114 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-52jpf" Dec 01 15:40:30 crc kubenswrapper[4739]: I1201 15:40:30.368177 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-52jpf" Dec 01 15:40:30 crc kubenswrapper[4739]: I1201 15:40:30.424644 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-52jpf" Dec 01 15:40:30 crc kubenswrapper[4739]: I1201 15:40:30.497013 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20f7f236-e00a-4002-9056-5099d2a6e485" path="/var/lib/kubelet/pods/20f7f236-e00a-4002-9056-5099d2a6e485/volumes" Dec 01 15:40:30 crc kubenswrapper[4739]: I1201 15:40:30.498319 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" event={"ID":"5433eac9-61a8-47e5-ba62-c9e09bcfe50b","Type":"ContainerStarted","Data":"5bbcf05ab3d2c138355b716707d9c9b0c1527c9c5c0460a1f191fbbefad9827f"} Dec 01 15:40:30 crc kubenswrapper[4739]: I1201 15:40:30.548130 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-52jpf" Dec 01 15:40:30 crc kubenswrapper[4739]: I1201 15:40:30.563857 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 15:40:31 crc kubenswrapper[4739]: I1201 15:40:31.488318 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" event={"ID":"5433eac9-61a8-47e5-ba62-c9e09bcfe50b","Type":"ContainerStarted","Data":"93f6b5715856196f5b8ac8cd03c8d5063949882cbfacc411e61ef0f2f50474e4"} Dec 01 15:40:31 crc kubenswrapper[4739]: I1201 15:40:31.516213 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" podStartSLOduration=4.516178103 podStartE2EDuration="4.516178103s" podCreationTimestamp="2025-12-01 15:40:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:40:31.515051888 +0000 UTC m=+333.340798012" watchObservedRunningTime="2025-12-01 15:40:31.516178103 +0000 UTC m=+333.341924237" Dec 01 15:40:32 crc kubenswrapper[4739]: I1201 15:40:32.495288 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:32 crc kubenswrapper[4739]: I1201 15:40:32.502289 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5598bcd6b-dsbdm" Dec 01 15:40:38 crc kubenswrapper[4739]: I1201 15:40:38.066246 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:40:38 crc kubenswrapper[4739]: I1201 15:40:38.119642 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:40:39 crc kubenswrapper[4739]: I1201 15:40:39.622309 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:40:39 crc kubenswrapper[4739]: I1201 15:40:39.622674 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:40:41 crc kubenswrapper[4739]: I1201 15:40:41.140115 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-bzzg4" Dec 01 15:40:41 crc kubenswrapper[4739]: I1201 15:40:41.210061 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kqk8p"] Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.270676 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" podUID="54c65dfa-5787-4f1f-979c-9e5daf517650" containerName="registry" containerID="cri-o://abfc89ffa169a3bf567891f19a5722c187642715f1a509b4706adf47c39630a7" gracePeriod=30 Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.722312 4739 generic.go:334] "Generic (PLEG): container finished" podID="54c65dfa-5787-4f1f-979c-9e5daf517650" containerID="abfc89ffa169a3bf567891f19a5722c187642715f1a509b4706adf47c39630a7" exitCode=0 Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.722488 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" event={"ID":"54c65dfa-5787-4f1f-979c-9e5daf517650","Type":"ContainerDied","Data":"abfc89ffa169a3bf567891f19a5722c187642715f1a509b4706adf47c39630a7"} Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.813038 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.917746 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"54c65dfa-5787-4f1f-979c-9e5daf517650\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.917799 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/54c65dfa-5787-4f1f-979c-9e5daf517650-installation-pull-secrets\") pod \"54c65dfa-5787-4f1f-979c-9e5daf517650\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.917822 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26wk4\" (UniqueName: \"kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-kube-api-access-26wk4\") pod \"54c65dfa-5787-4f1f-979c-9e5daf517650\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.917846 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-bound-sa-token\") pod \"54c65dfa-5787-4f1f-979c-9e5daf517650\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.917905 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/54c65dfa-5787-4f1f-979c-9e5daf517650-registry-certificates\") pod \"54c65dfa-5787-4f1f-979c-9e5daf517650\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.917953 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54c65dfa-5787-4f1f-979c-9e5daf517650-trusted-ca\") pod \"54c65dfa-5787-4f1f-979c-9e5daf517650\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.917977 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-registry-tls\") pod \"54c65dfa-5787-4f1f-979c-9e5daf517650\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.918000 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/54c65dfa-5787-4f1f-979c-9e5daf517650-ca-trust-extracted\") pod \"54c65dfa-5787-4f1f-979c-9e5daf517650\" (UID: \"54c65dfa-5787-4f1f-979c-9e5daf517650\") " Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.919790 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54c65dfa-5787-4f1f-979c-9e5daf517650-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "54c65dfa-5787-4f1f-979c-9e5daf517650" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.919811 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54c65dfa-5787-4f1f-979c-9e5daf517650-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "54c65dfa-5787-4f1f-979c-9e5daf517650" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.925101 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54c65dfa-5787-4f1f-979c-9e5daf517650-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "54c65dfa-5787-4f1f-979c-9e5daf517650" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.925122 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "54c65dfa-5787-4f1f-979c-9e5daf517650" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.925249 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-kube-api-access-26wk4" (OuterVolumeSpecName: "kube-api-access-26wk4") pod "54c65dfa-5787-4f1f-979c-9e5daf517650" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650"). InnerVolumeSpecName "kube-api-access-26wk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.927452 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "54c65dfa-5787-4f1f-979c-9e5daf517650" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.935865 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54c65dfa-5787-4f1f-979c-9e5daf517650-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "54c65dfa-5787-4f1f-979c-9e5daf517650" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:41:06 crc kubenswrapper[4739]: I1201 15:41:06.941688 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "54c65dfa-5787-4f1f-979c-9e5daf517650" (UID: "54c65dfa-5787-4f1f-979c-9e5daf517650"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 15:41:07 crc kubenswrapper[4739]: I1201 15:41:07.019502 4739 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54c65dfa-5787-4f1f-979c-9e5daf517650-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:41:07 crc kubenswrapper[4739]: I1201 15:41:07.019557 4739 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 15:41:07 crc kubenswrapper[4739]: I1201 15:41:07.019578 4739 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/54c65dfa-5787-4f1f-979c-9e5daf517650-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 15:41:07 crc kubenswrapper[4739]: I1201 15:41:07.019599 4739 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/54c65dfa-5787-4f1f-979c-9e5daf517650-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 15:41:07 crc kubenswrapper[4739]: I1201 15:41:07.019617 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26wk4\" (UniqueName: \"kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-kube-api-access-26wk4\") on node \"crc\" DevicePath \"\"" Dec 01 15:41:07 crc kubenswrapper[4739]: I1201 15:41:07.019636 4739 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/54c65dfa-5787-4f1f-979c-9e5daf517650-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 15:41:07 crc kubenswrapper[4739]: I1201 15:41:07.019657 4739 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/54c65dfa-5787-4f1f-979c-9e5daf517650-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 15:41:07 crc kubenswrapper[4739]: I1201 15:41:07.734715 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" event={"ID":"54c65dfa-5787-4f1f-979c-9e5daf517650","Type":"ContainerDied","Data":"056db4f59ea94ab8e09bedd015c01ec17b0faf85145d15e311d3baec0290f12c"} Dec 01 15:41:07 crc kubenswrapper[4739]: I1201 15:41:07.734812 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kqk8p" Dec 01 15:41:07 crc kubenswrapper[4739]: I1201 15:41:07.736459 4739 scope.go:117] "RemoveContainer" containerID="abfc89ffa169a3bf567891f19a5722c187642715f1a509b4706adf47c39630a7" Dec 01 15:41:07 crc kubenswrapper[4739]: I1201 15:41:07.782969 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kqk8p"] Dec 01 15:41:07 crc kubenswrapper[4739]: I1201 15:41:07.784917 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kqk8p"] Dec 01 15:41:08 crc kubenswrapper[4739]: I1201 15:41:08.496349 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54c65dfa-5787-4f1f-979c-9e5daf517650" path="/var/lib/kubelet/pods/54c65dfa-5787-4f1f-979c-9e5daf517650/volumes" Dec 01 15:41:09 crc kubenswrapper[4739]: I1201 15:41:09.621964 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:41:09 crc kubenswrapper[4739]: I1201 15:41:09.622067 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:41:39 crc kubenswrapper[4739]: I1201 15:41:39.621817 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:41:39 crc kubenswrapper[4739]: I1201 15:41:39.622507 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:41:39 crc kubenswrapper[4739]: I1201 15:41:39.622578 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:41:39 crc kubenswrapper[4739]: I1201 15:41:39.623618 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"682dc7ae590b7ec86c5c13fb0b344953a1f7c4e927c8f0da0655d437183e9a17"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:41:39 crc kubenswrapper[4739]: I1201 15:41:39.623717 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://682dc7ae590b7ec86c5c13fb0b344953a1f7c4e927c8f0da0655d437183e9a17" gracePeriod=600 Dec 01 15:41:39 crc kubenswrapper[4739]: I1201 15:41:39.946014 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="682dc7ae590b7ec86c5c13fb0b344953a1f7c4e927c8f0da0655d437183e9a17" exitCode=0 Dec 01 15:41:39 crc kubenswrapper[4739]: I1201 15:41:39.946071 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"682dc7ae590b7ec86c5c13fb0b344953a1f7c4e927c8f0da0655d437183e9a17"} Dec 01 15:41:39 crc kubenswrapper[4739]: I1201 15:41:39.947076 4739 scope.go:117] "RemoveContainer" containerID="924b0ce5d5d2c41463cec030556b2984eec58a0718d8622a729cf366339faa0e" Dec 01 15:41:40 crc kubenswrapper[4739]: I1201 15:41:40.954644 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"d55c29fa9b125ad4bffaf3cb1c8640b7606b8326b206fa44f12ce15e676ba1b1"} Dec 01 15:43:39 crc kubenswrapper[4739]: I1201 15:43:39.621573 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:43:39 crc kubenswrapper[4739]: I1201 15:43:39.624134 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:43:58 crc kubenswrapper[4739]: I1201 15:43:58.773487 4739 scope.go:117] "RemoveContainer" containerID="8b20bbc5f49dd9b0c6990e0a0e7a8c0bba83d323f9bfdebcc08405b29ffc3d7f" Dec 01 15:43:58 crc kubenswrapper[4739]: I1201 15:43:58.809064 4739 scope.go:117] "RemoveContainer" containerID="79eee8a23eb277a3572615023fd650c2e620e91a63e18bfeea078a2d2b855155" Dec 01 15:43:58 crc kubenswrapper[4739]: I1201 15:43:58.844926 4739 scope.go:117] "RemoveContainer" containerID="690f591b30eda2517b85c7ddcd303f3816a2bd6f20840dc26c61ff274a8504ad" Dec 01 15:44:09 crc kubenswrapper[4739]: I1201 15:44:09.621975 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:44:09 crc kubenswrapper[4739]: I1201 15:44:09.622706 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:44:39 crc kubenswrapper[4739]: I1201 15:44:39.622391 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:44:39 crc kubenswrapper[4739]: I1201 15:44:39.622973 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:44:39 crc kubenswrapper[4739]: I1201 15:44:39.623034 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:44:39 crc kubenswrapper[4739]: I1201 15:44:39.623769 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d55c29fa9b125ad4bffaf3cb1c8640b7606b8326b206fa44f12ce15e676ba1b1"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:44:39 crc kubenswrapper[4739]: I1201 15:44:39.623865 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://d55c29fa9b125ad4bffaf3cb1c8640b7606b8326b206fa44f12ce15e676ba1b1" gracePeriod=600 Dec 01 15:44:40 crc kubenswrapper[4739]: I1201 15:44:40.452460 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="d55c29fa9b125ad4bffaf3cb1c8640b7606b8326b206fa44f12ce15e676ba1b1" exitCode=0 Dec 01 15:44:40 crc kubenswrapper[4739]: I1201 15:44:40.452557 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"d55c29fa9b125ad4bffaf3cb1c8640b7606b8326b206fa44f12ce15e676ba1b1"} Dec 01 15:44:40 crc kubenswrapper[4739]: I1201 15:44:40.453320 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"e3830d5cb4e555123caab94e2aaa75fd65f71b1d428fbec87bdc6d807609b113"} Dec 01 15:44:40 crc kubenswrapper[4739]: I1201 15:44:40.453355 4739 scope.go:117] "RemoveContainer" containerID="682dc7ae590b7ec86c5c13fb0b344953a1f7c4e927c8f0da0655d437183e9a17" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.238087 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5"] Dec 01 15:45:00 crc kubenswrapper[4739]: E1201 15:45:00.238923 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54c65dfa-5787-4f1f-979c-9e5daf517650" containerName="registry" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.238938 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="54c65dfa-5787-4f1f-979c-9e5daf517650" containerName="registry" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.239057 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="54c65dfa-5787-4f1f-979c-9e5daf517650" containerName="registry" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.239490 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.242823 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.242928 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.262233 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5"] Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.363088 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57gnx\" (UniqueName: \"kubernetes.io/projected/e6950571-e8f3-4e9c-b2e8-05a806722f8b-kube-api-access-57gnx\") pod \"collect-profiles-29410065-nq4j5\" (UID: \"e6950571-e8f3-4e9c-b2e8-05a806722f8b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.363290 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6950571-e8f3-4e9c-b2e8-05a806722f8b-config-volume\") pod \"collect-profiles-29410065-nq4j5\" (UID: \"e6950571-e8f3-4e9c-b2e8-05a806722f8b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.363329 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6950571-e8f3-4e9c-b2e8-05a806722f8b-secret-volume\") pod \"collect-profiles-29410065-nq4j5\" (UID: \"e6950571-e8f3-4e9c-b2e8-05a806722f8b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.464393 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57gnx\" (UniqueName: \"kubernetes.io/projected/e6950571-e8f3-4e9c-b2e8-05a806722f8b-kube-api-access-57gnx\") pod \"collect-profiles-29410065-nq4j5\" (UID: \"e6950571-e8f3-4e9c-b2e8-05a806722f8b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.464611 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6950571-e8f3-4e9c-b2e8-05a806722f8b-config-volume\") pod \"collect-profiles-29410065-nq4j5\" (UID: \"e6950571-e8f3-4e9c-b2e8-05a806722f8b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.464650 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6950571-e8f3-4e9c-b2e8-05a806722f8b-secret-volume\") pod \"collect-profiles-29410065-nq4j5\" (UID: \"e6950571-e8f3-4e9c-b2e8-05a806722f8b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.466140 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6950571-e8f3-4e9c-b2e8-05a806722f8b-config-volume\") pod \"collect-profiles-29410065-nq4j5\" (UID: \"e6950571-e8f3-4e9c-b2e8-05a806722f8b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.480947 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6950571-e8f3-4e9c-b2e8-05a806722f8b-secret-volume\") pod \"collect-profiles-29410065-nq4j5\" (UID: \"e6950571-e8f3-4e9c-b2e8-05a806722f8b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.500771 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57gnx\" (UniqueName: \"kubernetes.io/projected/e6950571-e8f3-4e9c-b2e8-05a806722f8b-kube-api-access-57gnx\") pod \"collect-profiles-29410065-nq4j5\" (UID: \"e6950571-e8f3-4e9c-b2e8-05a806722f8b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.556015 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" Dec 01 15:45:00 crc kubenswrapper[4739]: I1201 15:45:00.981647 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5"] Dec 01 15:45:00 crc kubenswrapper[4739]: W1201 15:45:00.989207 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6950571_e8f3_4e9c_b2e8_05a806722f8b.slice/crio-a7b1ecab99f6b6fc90885390ca2da22da630ee2ab44119f65881718b7869a90a WatchSource:0}: Error finding container a7b1ecab99f6b6fc90885390ca2da22da630ee2ab44119f65881718b7869a90a: Status 404 returned error can't find the container with id a7b1ecab99f6b6fc90885390ca2da22da630ee2ab44119f65881718b7869a90a Dec 01 15:45:01 crc kubenswrapper[4739]: I1201 15:45:01.594906 4739 generic.go:334] "Generic (PLEG): container finished" podID="e6950571-e8f3-4e9c-b2e8-05a806722f8b" containerID="11fa0ea5b5cf8b38a4664f369c5f50fb93ddfa0697829736c99ba7bdc953acef" exitCode=0 Dec 01 15:45:01 crc kubenswrapper[4739]: I1201 15:45:01.594974 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" event={"ID":"e6950571-e8f3-4e9c-b2e8-05a806722f8b","Type":"ContainerDied","Data":"11fa0ea5b5cf8b38a4664f369c5f50fb93ddfa0697829736c99ba7bdc953acef"} Dec 01 15:45:01 crc kubenswrapper[4739]: I1201 15:45:01.596894 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" event={"ID":"e6950571-e8f3-4e9c-b2e8-05a806722f8b","Type":"ContainerStarted","Data":"a7b1ecab99f6b6fc90885390ca2da22da630ee2ab44119f65881718b7869a90a"} Dec 01 15:45:02 crc kubenswrapper[4739]: I1201 15:45:02.916956 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" Dec 01 15:45:03 crc kubenswrapper[4739]: I1201 15:45:03.001392 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57gnx\" (UniqueName: \"kubernetes.io/projected/e6950571-e8f3-4e9c-b2e8-05a806722f8b-kube-api-access-57gnx\") pod \"e6950571-e8f3-4e9c-b2e8-05a806722f8b\" (UID: \"e6950571-e8f3-4e9c-b2e8-05a806722f8b\") " Dec 01 15:45:03 crc kubenswrapper[4739]: I1201 15:45:03.001554 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6950571-e8f3-4e9c-b2e8-05a806722f8b-secret-volume\") pod \"e6950571-e8f3-4e9c-b2e8-05a806722f8b\" (UID: \"e6950571-e8f3-4e9c-b2e8-05a806722f8b\") " Dec 01 15:45:03 crc kubenswrapper[4739]: I1201 15:45:03.002980 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6950571-e8f3-4e9c-b2e8-05a806722f8b-config-volume\") pod \"e6950571-e8f3-4e9c-b2e8-05a806722f8b\" (UID: \"e6950571-e8f3-4e9c-b2e8-05a806722f8b\") " Dec 01 15:45:03 crc kubenswrapper[4739]: I1201 15:45:03.003903 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6950571-e8f3-4e9c-b2e8-05a806722f8b-config-volume" (OuterVolumeSpecName: "config-volume") pod "e6950571-e8f3-4e9c-b2e8-05a806722f8b" (UID: "e6950571-e8f3-4e9c-b2e8-05a806722f8b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:45:03 crc kubenswrapper[4739]: I1201 15:45:03.010149 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6950571-e8f3-4e9c-b2e8-05a806722f8b-kube-api-access-57gnx" (OuterVolumeSpecName: "kube-api-access-57gnx") pod "e6950571-e8f3-4e9c-b2e8-05a806722f8b" (UID: "e6950571-e8f3-4e9c-b2e8-05a806722f8b"). InnerVolumeSpecName "kube-api-access-57gnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:45:03 crc kubenswrapper[4739]: I1201 15:45:03.010402 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6950571-e8f3-4e9c-b2e8-05a806722f8b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e6950571-e8f3-4e9c-b2e8-05a806722f8b" (UID: "e6950571-e8f3-4e9c-b2e8-05a806722f8b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:45:03 crc kubenswrapper[4739]: I1201 15:45:03.104569 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57gnx\" (UniqueName: \"kubernetes.io/projected/e6950571-e8f3-4e9c-b2e8-05a806722f8b-kube-api-access-57gnx\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:03 crc kubenswrapper[4739]: I1201 15:45:03.104818 4739 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6950571-e8f3-4e9c-b2e8-05a806722f8b-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:03 crc kubenswrapper[4739]: I1201 15:45:03.104907 4739 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6950571-e8f3-4e9c-b2e8-05a806722f8b-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:03 crc kubenswrapper[4739]: I1201 15:45:03.613841 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" event={"ID":"e6950571-e8f3-4e9c-b2e8-05a806722f8b","Type":"ContainerDied","Data":"a7b1ecab99f6b6fc90885390ca2da22da630ee2ab44119f65881718b7869a90a"} Dec 01 15:45:03 crc kubenswrapper[4739]: I1201 15:45:03.613898 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7b1ecab99f6b6fc90885390ca2da22da630ee2ab44119f65881718b7869a90a" Dec 01 15:45:03 crc kubenswrapper[4739]: I1201 15:45:03.613937 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5" Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.758218 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qp8rs"] Dec 01 15:45:24 crc kubenswrapper[4739]: E1201 15:45:24.759062 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6950571-e8f3-4e9c-b2e8-05a806722f8b" containerName="collect-profiles" Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.759081 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6950571-e8f3-4e9c-b2e8-05a806722f8b" containerName="collect-profiles" Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.759227 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6950571-e8f3-4e9c-b2e8-05a806722f8b" containerName="collect-profiles" Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.761307 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qp8rs" Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.764126 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.764600 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.766563 4739 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-dmkl4" Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.769375 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qp8rs"] Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.776606 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-d4qnq"] Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.777366 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-d4qnq" Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.779982 4739 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-285sd" Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.790934 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-fm6tr"] Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.791894 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-fm6tr" Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.795291 4739 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-xlt5h" Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.803201 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-d4qnq"] Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.809566 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-fm6tr"] Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.911540 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nktd\" (UniqueName: \"kubernetes.io/projected/56175826-6fcb-4142-8716-e4fb8664a8b8-kube-api-access-9nktd\") pod \"cert-manager-webhook-5655c58dd6-fm6tr\" (UID: \"56175826-6fcb-4142-8716-e4fb8664a8b8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-fm6tr" Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.911685 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44dfz\" (UniqueName: \"kubernetes.io/projected/04583995-14a7-4263-ac42-3628b4e5760c-kube-api-access-44dfz\") pod \"cert-manager-cainjector-7f985d654d-qp8rs\" (UID: \"04583995-14a7-4263-ac42-3628b4e5760c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qp8rs" Dec 01 15:45:24 crc kubenswrapper[4739]: I1201 15:45:24.911726 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7qlr\" (UniqueName: \"kubernetes.io/projected/a3382af1-0e5e-41cc-9446-8f0a66d741cb-kube-api-access-m7qlr\") pod \"cert-manager-5b446d88c5-d4qnq\" (UID: \"a3382af1-0e5e-41cc-9446-8f0a66d741cb\") " pod="cert-manager/cert-manager-5b446d88c5-d4qnq" Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.012756 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nktd\" (UniqueName: \"kubernetes.io/projected/56175826-6fcb-4142-8716-e4fb8664a8b8-kube-api-access-9nktd\") pod \"cert-manager-webhook-5655c58dd6-fm6tr\" (UID: \"56175826-6fcb-4142-8716-e4fb8664a8b8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-fm6tr" Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.012893 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44dfz\" (UniqueName: \"kubernetes.io/projected/04583995-14a7-4263-ac42-3628b4e5760c-kube-api-access-44dfz\") pod \"cert-manager-cainjector-7f985d654d-qp8rs\" (UID: \"04583995-14a7-4263-ac42-3628b4e5760c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qp8rs" Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.012929 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7qlr\" (UniqueName: \"kubernetes.io/projected/a3382af1-0e5e-41cc-9446-8f0a66d741cb-kube-api-access-m7qlr\") pod \"cert-manager-5b446d88c5-d4qnq\" (UID: \"a3382af1-0e5e-41cc-9446-8f0a66d741cb\") " pod="cert-manager/cert-manager-5b446d88c5-d4qnq" Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.031732 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7qlr\" (UniqueName: \"kubernetes.io/projected/a3382af1-0e5e-41cc-9446-8f0a66d741cb-kube-api-access-m7qlr\") pod \"cert-manager-5b446d88c5-d4qnq\" (UID: \"a3382af1-0e5e-41cc-9446-8f0a66d741cb\") " pod="cert-manager/cert-manager-5b446d88c5-d4qnq" Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.031966 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nktd\" (UniqueName: \"kubernetes.io/projected/56175826-6fcb-4142-8716-e4fb8664a8b8-kube-api-access-9nktd\") pod \"cert-manager-webhook-5655c58dd6-fm6tr\" (UID: \"56175826-6fcb-4142-8716-e4fb8664a8b8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-fm6tr" Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.040352 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44dfz\" (UniqueName: \"kubernetes.io/projected/04583995-14a7-4263-ac42-3628b4e5760c-kube-api-access-44dfz\") pod \"cert-manager-cainjector-7f985d654d-qp8rs\" (UID: \"04583995-14a7-4263-ac42-3628b4e5760c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qp8rs" Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.083255 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qp8rs" Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.094006 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-d4qnq" Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.103496 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-fm6tr" Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.413833 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-d4qnq"] Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.422907 4739 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.555998 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qp8rs"] Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.559813 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-fm6tr"] Dec 01 15:45:25 crc kubenswrapper[4739]: W1201 15:45:25.563758 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04583995_14a7_4263_ac42_3628b4e5760c.slice/crio-69c68cdb4bc8eb340eb2c6146d211a2a836b21f2f8a612567923e69c2d3708b8 WatchSource:0}: Error finding container 69c68cdb4bc8eb340eb2c6146d211a2a836b21f2f8a612567923e69c2d3708b8: Status 404 returned error can't find the container with id 69c68cdb4bc8eb340eb2c6146d211a2a836b21f2f8a612567923e69c2d3708b8 Dec 01 15:45:25 crc kubenswrapper[4739]: W1201 15:45:25.564360 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56175826_6fcb_4142_8716_e4fb8664a8b8.slice/crio-233f11f5875c85b83145c958ecdd7cea5ca985ac772ace13cd7608705c78cff1 WatchSource:0}: Error finding container 233f11f5875c85b83145c958ecdd7cea5ca985ac772ace13cd7608705c78cff1: Status 404 returned error can't find the container with id 233f11f5875c85b83145c958ecdd7cea5ca985ac772ace13cd7608705c78cff1 Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.755174 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-fm6tr" event={"ID":"56175826-6fcb-4142-8716-e4fb8664a8b8","Type":"ContainerStarted","Data":"233f11f5875c85b83145c958ecdd7cea5ca985ac772ace13cd7608705c78cff1"} Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.756436 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-d4qnq" event={"ID":"a3382af1-0e5e-41cc-9446-8f0a66d741cb","Type":"ContainerStarted","Data":"c56f8c8a6f59752c5388a82aec12d5194302c07aaf20a7d1afc5229e1b69aae2"} Dec 01 15:45:25 crc kubenswrapper[4739]: I1201 15:45:25.757684 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qp8rs" event={"ID":"04583995-14a7-4263-ac42-3628b4e5760c","Type":"ContainerStarted","Data":"69c68cdb4bc8eb340eb2c6146d211a2a836b21f2f8a612567923e69c2d3708b8"} Dec 01 15:45:28 crc kubenswrapper[4739]: I1201 15:45:28.776382 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-fm6tr" event={"ID":"56175826-6fcb-4142-8716-e4fb8664a8b8","Type":"ContainerStarted","Data":"e0d7c03aa71fd916ba341a4c2627bd9d16d857a788f4c98fda545b0100f8798f"} Dec 01 15:45:28 crc kubenswrapper[4739]: I1201 15:45:28.777001 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-fm6tr" Dec 01 15:45:28 crc kubenswrapper[4739]: I1201 15:45:28.787747 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-d4qnq" event={"ID":"a3382af1-0e5e-41cc-9446-8f0a66d741cb","Type":"ContainerStarted","Data":"64891dd7d78894bbb6271c440c7e554c13bfac1d6002f6e9fe01bec15da39188"} Dec 01 15:45:28 crc kubenswrapper[4739]: I1201 15:45:28.803401 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-fm6tr" podStartSLOduration=2.18730348 podStartE2EDuration="4.803382649s" podCreationTimestamp="2025-12-01 15:45:24 +0000 UTC" firstStartedPulling="2025-12-01 15:45:25.566789329 +0000 UTC m=+627.392535433" lastFinishedPulling="2025-12-01 15:45:28.182868468 +0000 UTC m=+630.008614602" observedRunningTime="2025-12-01 15:45:28.799815275 +0000 UTC m=+630.625561369" watchObservedRunningTime="2025-12-01 15:45:28.803382649 +0000 UTC m=+630.629128743" Dec 01 15:45:28 crc kubenswrapper[4739]: I1201 15:45:28.816506 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-d4qnq" podStartSLOduration=2.098912824 podStartE2EDuration="4.816489774s" podCreationTimestamp="2025-12-01 15:45:24 +0000 UTC" firstStartedPulling="2025-12-01 15:45:25.422664225 +0000 UTC m=+627.248410319" lastFinishedPulling="2025-12-01 15:45:28.140241175 +0000 UTC m=+629.965987269" observedRunningTime="2025-12-01 15:45:28.814023996 +0000 UTC m=+630.639770140" watchObservedRunningTime="2025-12-01 15:45:28.816489774 +0000 UTC m=+630.642235868" Dec 01 15:45:29 crc kubenswrapper[4739]: I1201 15:45:29.794187 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qp8rs" event={"ID":"04583995-14a7-4263-ac42-3628b4e5760c","Type":"ContainerStarted","Data":"7537dfc4390e03142537de945a1ef80d0c4ef46a98902e5a9b98eceff3b0dbe3"} Dec 01 15:45:29 crc kubenswrapper[4739]: I1201 15:45:29.861648 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-qp8rs" podStartSLOduration=2.410384929 podStartE2EDuration="5.861632981s" podCreationTimestamp="2025-12-01 15:45:24 +0000 UTC" firstStartedPulling="2025-12-01 15:45:25.565372954 +0000 UTC m=+627.391119048" lastFinishedPulling="2025-12-01 15:45:29.016621006 +0000 UTC m=+630.842367100" observedRunningTime="2025-12-01 15:45:29.859131002 +0000 UTC m=+631.684877096" watchObservedRunningTime="2025-12-01 15:45:29.861632981 +0000 UTC m=+631.687379075" Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.107227 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-fm6tr" Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.356940 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dlvkg"] Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.357307 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="northd" containerID="cri-o://d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018" gracePeriod=30 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.357459 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="sbdb" containerID="cri-o://13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f" gracePeriod=30 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.357492 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="nbdb" containerID="cri-o://63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e" gracePeriod=30 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.357659 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="kube-rbac-proxy-node" containerID="cri-o://bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24" gracePeriod=30 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.357696 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706" gracePeriod=30 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.357732 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovn-acl-logging" containerID="cri-o://a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609" gracePeriod=30 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.357772 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovn-controller" containerID="cri-o://76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465" gracePeriod=30 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.390705 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" containerID="cri-o://edcd593134ac40a8f73daf48b1614b2fe6e7d0f598dd6cd8b741934e8dfbb1dc" gracePeriod=30 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.833901 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovnkube-controller/3.log" Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.836992 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovn-acl-logging/0.log" Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.837802 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovn-controller/0.log" Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838327 4739 generic.go:334] "Generic (PLEG): container finished" podID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerID="edcd593134ac40a8f73daf48b1614b2fe6e7d0f598dd6cd8b741934e8dfbb1dc" exitCode=0 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838403 4739 generic.go:334] "Generic (PLEG): container finished" podID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerID="13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f" exitCode=0 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838448 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerDied","Data":"edcd593134ac40a8f73daf48b1614b2fe6e7d0f598dd6cd8b741934e8dfbb1dc"} Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838473 4739 generic.go:334] "Generic (PLEG): container finished" podID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerID="63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e" exitCode=0 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838495 4739 generic.go:334] "Generic (PLEG): container finished" podID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerID="d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018" exitCode=0 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838512 4739 generic.go:334] "Generic (PLEG): container finished" podID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerID="9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706" exitCode=0 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838531 4739 generic.go:334] "Generic (PLEG): container finished" podID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerID="bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24" exitCode=0 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838537 4739 scope.go:117] "RemoveContainer" containerID="e5117f28a7ce17e75a635e61cc3f1cbba0419c8a01b53c91a2ecf63055585fbb" Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838550 4739 generic.go:334] "Generic (PLEG): container finished" podID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerID="a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609" exitCode=143 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838569 4739 generic.go:334] "Generic (PLEG): container finished" podID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerID="76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465" exitCode=143 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838516 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerDied","Data":"13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f"} Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838692 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerDied","Data":"63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e"} Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838729 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerDied","Data":"d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018"} Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838759 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerDied","Data":"9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706"} Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838784 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerDied","Data":"bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24"} Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838811 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerDied","Data":"a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609"} Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.838834 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerDied","Data":"76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465"} Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.841595 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nt6tv_341a0b94-e13d-45cb-8f1c-2c1d1f101272/kube-multus/1.log" Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.842310 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nt6tv_341a0b94-e13d-45cb-8f1c-2c1d1f101272/kube-multus/0.log" Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.842379 4739 generic.go:334] "Generic (PLEG): container finished" podID="341a0b94-e13d-45cb-8f1c-2c1d1f101272" containerID="b580da066f2e8e584d58d7c8982dbc4accfa0bd4149637292b4e8550449452f8" exitCode=2 Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.842446 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nt6tv" event={"ID":"341a0b94-e13d-45cb-8f1c-2c1d1f101272","Type":"ContainerDied","Data":"b580da066f2e8e584d58d7c8982dbc4accfa0bd4149637292b4e8550449452f8"} Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.843129 4739 scope.go:117] "RemoveContainer" containerID="b580da066f2e8e584d58d7c8982dbc4accfa0bd4149637292b4e8550449452f8" Dec 01 15:45:35 crc kubenswrapper[4739]: E1201 15:45:35.843463 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-nt6tv_openshift-multus(341a0b94-e13d-45cb-8f1c-2c1d1f101272)\"" pod="openshift-multus/multus-nt6tv" podUID="341a0b94-e13d-45cb-8f1c-2c1d1f101272" Dec 01 15:45:35 crc kubenswrapper[4739]: I1201 15:45:35.859883 4739 scope.go:117] "RemoveContainer" containerID="13bc86e44229abbcce97d80f1e7b611270b15e4e41d37798ff4284a96ef9aa46" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.206191 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovn-acl-logging/0.log" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.206801 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovn-controller/0.log" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.207320 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.275958 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ml5qz"] Dec 01 15:45:36 crc kubenswrapper[4739]: E1201 15:45:36.276158 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276172 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: E1201 15:45:36.276183 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="northd" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276193 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="northd" Dec 01 15:45:36 crc kubenswrapper[4739]: E1201 15:45:36.276208 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276215 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: E1201 15:45:36.276227 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="sbdb" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276234 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="sbdb" Dec 01 15:45:36 crc kubenswrapper[4739]: E1201 15:45:36.276244 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="kube-rbac-proxy-node" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276251 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="kube-rbac-proxy-node" Dec 01 15:45:36 crc kubenswrapper[4739]: E1201 15:45:36.276261 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="nbdb" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276267 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="nbdb" Dec 01 15:45:36 crc kubenswrapper[4739]: E1201 15:45:36.276275 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovn-acl-logging" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276282 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovn-acl-logging" Dec 01 15:45:36 crc kubenswrapper[4739]: E1201 15:45:36.276293 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276299 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: E1201 15:45:36.276306 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276313 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: E1201 15:45:36.276325 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276333 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 15:45:36 crc kubenswrapper[4739]: E1201 15:45:36.276344 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276351 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: E1201 15:45:36.276363 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovn-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276371 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovn-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: E1201 15:45:36.276384 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="kubecfg-setup" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276391 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="kubecfg-setup" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276546 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276560 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="sbdb" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276566 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="northd" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276578 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276586 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovn-acl-logging" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276596 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="nbdb" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276607 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276632 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276639 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="kube-rbac-proxy-node" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276647 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276652 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovnkube-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.276660 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" containerName="ovn-controller" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.279665 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364217 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-ovn\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364283 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovnkube-config\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364319 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovnkube-script-lib\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364356 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovn-node-metrics-cert\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364393 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-slash\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364447 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvdpq\" (UniqueName: \"kubernetes.io/projected/43a8da1c-7c68-4af9-8b57-817e55d3b875-kube-api-access-rvdpq\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364470 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-cni-bin\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364488 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-kubelet\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364510 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-var-lib-openvswitch\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364555 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-systemd-units\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364573 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-run-netns\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364600 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-log-socket\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364632 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-cni-netd\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364652 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-openvswitch\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364679 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-run-ovn-kubernetes\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364702 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-var-lib-cni-networks-ovn-kubernetes\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364742 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-systemd\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364765 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-env-overrides\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364791 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-node-log\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364817 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-etc-openvswitch\") pod \"43a8da1c-7c68-4af9-8b57-817e55d3b875\" (UID: \"43a8da1c-7c68-4af9-8b57-817e55d3b875\") " Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.364967 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a411094b-a18c-497b-a82c-678e709d394d-ovn-node-metrics-cert\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365003 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365007 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-slash\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365046 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-etc-openvswitch\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365071 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-run-ovn\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365084 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-log-socket" (OuterVolumeSpecName: "log-socket") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365092 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-run-ovn-kubernetes\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365096 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365127 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-cni-netd\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365112 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365145 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365148 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365126 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365171 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365170 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365177 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-node-log" (OuterVolumeSpecName: "node-log") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365212 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365224 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365267 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-slash" (OuterVolumeSpecName: "host-slash") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365308 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a411094b-a18c-497b-a82c-678e709d394d-env-overrides\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365347 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vskbr\" (UniqueName: \"kubernetes.io/projected/a411094b-a18c-497b-a82c-678e709d394d-kube-api-access-vskbr\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365385 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-log-socket\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365471 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365503 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365515 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a411094b-a18c-497b-a82c-678e709d394d-ovnkube-config\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365550 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a411094b-a18c-497b-a82c-678e709d394d-ovnkube-script-lib\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365586 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-node-log\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365648 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-run-netns\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365677 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-cni-bin\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365715 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-run-openvswitch\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365757 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-var-lib-openvswitch\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365788 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-systemd-units\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365827 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-kubelet\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365856 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-run-systemd\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365913 4739 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365927 4739 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365940 4739 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365952 4739 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365963 4739 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-node-log\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365974 4739 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.365988 4739 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.366000 4739 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-slash\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.366057 4739 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.366068 4739 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.366079 4739 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.366091 4739 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.366103 4739 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.366114 4739 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-log-socket\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.366547 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.366597 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.367027 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.370641 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.371038 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43a8da1c-7c68-4af9-8b57-817e55d3b875-kube-api-access-rvdpq" (OuterVolumeSpecName: "kube-api-access-rvdpq") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "kube-api-access-rvdpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.388835 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "43a8da1c-7c68-4af9-8b57-817e55d3b875" (UID: "43a8da1c-7c68-4af9-8b57-817e55d3b875"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467221 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467324 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a411094b-a18c-497b-a82c-678e709d394d-ovnkube-config\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467359 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a411094b-a18c-497b-a82c-678e709d394d-ovnkube-script-lib\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467357 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467393 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-node-log\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467495 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-run-netns\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467527 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-cni-bin\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467569 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-run-openvswitch\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467612 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-var-lib-openvswitch\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467614 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-run-netns\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467643 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-systemd-units\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467708 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-run-openvswitch\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467734 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-kubelet\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467745 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-var-lib-openvswitch\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467757 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-cni-bin\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467761 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-run-systemd\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467784 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-run-systemd\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467708 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-systemd-units\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467801 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-kubelet\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467844 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-node-log\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467868 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a411094b-a18c-497b-a82c-678e709d394d-ovn-node-metrics-cert\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467912 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-run-ovn\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467928 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-run-ovn-kubernetes\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467947 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-cni-netd\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467961 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-slash\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.467977 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-etc-openvswitch\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468008 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a411094b-a18c-497b-a82c-678e709d394d-env-overrides\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468030 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vskbr\" (UniqueName: \"kubernetes.io/projected/a411094b-a18c-497b-a82c-678e709d394d-kube-api-access-vskbr\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468055 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-log-socket\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468116 4739 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468126 4739 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/43a8da1c-7c68-4af9-8b57-817e55d3b875-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468138 4739 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468149 4739 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468157 4739 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/43a8da1c-7c68-4af9-8b57-817e55d3b875-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468168 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvdpq\" (UniqueName: \"kubernetes.io/projected/43a8da1c-7c68-4af9-8b57-817e55d3b875-kube-api-access-rvdpq\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468192 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-log-socket\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468214 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-run-ovn\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468234 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-run-ovn-kubernetes\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468254 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-cni-netd\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468271 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-host-slash\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468290 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a411094b-a18c-497b-a82c-678e709d394d-etc-openvswitch\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468572 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a411094b-a18c-497b-a82c-678e709d394d-ovnkube-config\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468708 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a411094b-a18c-497b-a82c-678e709d394d-env-overrides\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.468866 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a411094b-a18c-497b-a82c-678e709d394d-ovnkube-script-lib\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.470994 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a411094b-a18c-497b-a82c-678e709d394d-ovn-node-metrics-cert\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.497160 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vskbr\" (UniqueName: \"kubernetes.io/projected/a411094b-a18c-497b-a82c-678e709d394d-kube-api-access-vskbr\") pod \"ovnkube-node-ml5qz\" (UID: \"a411094b-a18c-497b-a82c-678e709d394d\") " pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.602906 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.859470 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" event={"ID":"a411094b-a18c-497b-a82c-678e709d394d","Type":"ContainerStarted","Data":"62fa38b52cc70c069319ec391f8b9622a69e6f3b15e56a30187d2478560a1574"} Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.878413 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovn-acl-logging/0.log" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.879360 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dlvkg_43a8da1c-7c68-4af9-8b57-817e55d3b875/ovn-controller/0.log" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.880246 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.880206 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dlvkg" event={"ID":"43a8da1c-7c68-4af9-8b57-817e55d3b875","Type":"ContainerDied","Data":"8f334e545ef5edead52a18e9b43acc2e71e4c4867917134a24881d8668f52997"} Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.880840 4739 scope.go:117] "RemoveContainer" containerID="edcd593134ac40a8f73daf48b1614b2fe6e7d0f598dd6cd8b741934e8dfbb1dc" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.883818 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nt6tv_341a0b94-e13d-45cb-8f1c-2c1d1f101272/kube-multus/1.log" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.922054 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dlvkg"] Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.922349 4739 scope.go:117] "RemoveContainer" containerID="13fd9b3f36db7c359b25fc6bffd6e2ddd64f1e94f629242e5b997403121a6b4f" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.931590 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dlvkg"] Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.936135 4739 scope.go:117] "RemoveContainer" containerID="63e13f282326382f351ae0e338fee9eb2e6e69c6e9290d09bec48548f902971e" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.953271 4739 scope.go:117] "RemoveContainer" containerID="d9e98c0fad13e6f906f6daa95254a215a6a52b5d71c40941eba12722d280f018" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.965652 4739 scope.go:117] "RemoveContainer" containerID="9ded6f1d057804af01d0eb1a1932b74f23608dd38a8ca58af2c0c5c710615706" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.981309 4739 scope.go:117] "RemoveContainer" containerID="bb1412edfbf1c93e9bac09f6b1984e8fb7f827115c907a27e06cac94b2e2ee24" Dec 01 15:45:36 crc kubenswrapper[4739]: I1201 15:45:36.998829 4739 scope.go:117] "RemoveContainer" containerID="a2e2083dd097500833b7553980dd00a892a88e838d0a03d21f7787c68c044609" Dec 01 15:45:37 crc kubenswrapper[4739]: I1201 15:45:37.010534 4739 scope.go:117] "RemoveContainer" containerID="76b522deb3b14ba1812d949d84ffd7b3c132aba55057e1b336c68df48bdb7465" Dec 01 15:45:37 crc kubenswrapper[4739]: I1201 15:45:37.026319 4739 scope.go:117] "RemoveContainer" containerID="e2220ed8cb513b428ed4bf8406f9d5251f7b778ab1fc59d2f54174bd763db55b" Dec 01 15:45:37 crc kubenswrapper[4739]: I1201 15:45:37.894693 4739 generic.go:334] "Generic (PLEG): container finished" podID="a411094b-a18c-497b-a82c-678e709d394d" containerID="eb587b34071162ccccffbe4b80974dd80f5dedb0af09b0f3e9fd09cf7e668585" exitCode=0 Dec 01 15:45:37 crc kubenswrapper[4739]: I1201 15:45:37.894799 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" event={"ID":"a411094b-a18c-497b-a82c-678e709d394d","Type":"ContainerDied","Data":"eb587b34071162ccccffbe4b80974dd80f5dedb0af09b0f3e9fd09cf7e668585"} Dec 01 15:45:38 crc kubenswrapper[4739]: I1201 15:45:38.484288 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43a8da1c-7c68-4af9-8b57-817e55d3b875" path="/var/lib/kubelet/pods/43a8da1c-7c68-4af9-8b57-817e55d3b875/volumes" Dec 01 15:45:38 crc kubenswrapper[4739]: I1201 15:45:38.909701 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" event={"ID":"a411094b-a18c-497b-a82c-678e709d394d","Type":"ContainerStarted","Data":"d2a00a6de81714ef2f86ab9b7c0012d19167aa2402f7cd50e3b4d8e6d63cc845"} Dec 01 15:45:38 crc kubenswrapper[4739]: I1201 15:45:38.911752 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" event={"ID":"a411094b-a18c-497b-a82c-678e709d394d","Type":"ContainerStarted","Data":"5bc96228eafbb51aaafd91c2170a07b60561534dde785b4042ceec68e5d95a02"} Dec 01 15:45:38 crc kubenswrapper[4739]: I1201 15:45:38.911912 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" event={"ID":"a411094b-a18c-497b-a82c-678e709d394d","Type":"ContainerStarted","Data":"4586db9b26ea390a599ad5219c54da1fd3ce7085cc2d7c3b02ae5a5147d30c2d"} Dec 01 15:45:38 crc kubenswrapper[4739]: I1201 15:45:38.912035 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" event={"ID":"a411094b-a18c-497b-a82c-678e709d394d","Type":"ContainerStarted","Data":"1b6a739d25e4aac3b84c4e4fdde17f100945a2f7855dae87b1f125e2b8dc1897"} Dec 01 15:45:38 crc kubenswrapper[4739]: I1201 15:45:38.912156 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" event={"ID":"a411094b-a18c-497b-a82c-678e709d394d","Type":"ContainerStarted","Data":"dc0901d853ae3383c13f1e79526c7009538a3b5ff3ac8a5ba572abe30a203210"} Dec 01 15:45:38 crc kubenswrapper[4739]: I1201 15:45:38.912266 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" event={"ID":"a411094b-a18c-497b-a82c-678e709d394d","Type":"ContainerStarted","Data":"4de6792572d8cd08a7afefda479b96138faafda7617c0c25d564bf6858999c40"} Dec 01 15:45:41 crc kubenswrapper[4739]: I1201 15:45:41.943194 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" event={"ID":"a411094b-a18c-497b-a82c-678e709d394d","Type":"ContainerStarted","Data":"cc0f005ae04c4ac51b2a106aa1cbbe9486727649071e91fd401eb5693553bbd3"} Dec 01 15:45:43 crc kubenswrapper[4739]: I1201 15:45:43.960030 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" event={"ID":"a411094b-a18c-497b-a82c-678e709d394d","Type":"ContainerStarted","Data":"0a459697360922d5e61e6b65d71c938ca632b3bed3113e2328362e1749f9297f"} Dec 01 15:45:43 crc kubenswrapper[4739]: I1201 15:45:43.960535 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:43 crc kubenswrapper[4739]: I1201 15:45:43.960572 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:43 crc kubenswrapper[4739]: I1201 15:45:43.960599 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:43 crc kubenswrapper[4739]: I1201 15:45:43.999343 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" podStartSLOduration=7.999312265 podStartE2EDuration="7.999312265s" podCreationTimestamp="2025-12-01 15:45:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:45:43.996798415 +0000 UTC m=+645.822544519" watchObservedRunningTime="2025-12-01 15:45:43.999312265 +0000 UTC m=+645.825058359" Dec 01 15:45:44 crc kubenswrapper[4739]: I1201 15:45:44.009277 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:44 crc kubenswrapper[4739]: I1201 15:45:44.015440 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:45:48 crc kubenswrapper[4739]: I1201 15:45:48.484931 4739 scope.go:117] "RemoveContainer" containerID="b580da066f2e8e584d58d7c8982dbc4accfa0bd4149637292b4e8550449452f8" Dec 01 15:45:48 crc kubenswrapper[4739]: I1201 15:45:48.995806 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nt6tv_341a0b94-e13d-45cb-8f1c-2c1d1f101272/kube-multus/1.log" Dec 01 15:45:48 crc kubenswrapper[4739]: I1201 15:45:48.996310 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nt6tv" event={"ID":"341a0b94-e13d-45cb-8f1c-2c1d1f101272","Type":"ContainerStarted","Data":"995b23431b7db5f902363c003a4c9ee2fcdc59850216bffd4aa576f1cf9014d0"} Dec 01 15:46:06 crc kubenswrapper[4739]: I1201 15:46:06.644475 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ml5qz" Dec 01 15:46:15 crc kubenswrapper[4739]: I1201 15:46:15.805957 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp"] Dec 01 15:46:15 crc kubenswrapper[4739]: I1201 15:46:15.809234 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" Dec 01 15:46:15 crc kubenswrapper[4739]: I1201 15:46:15.812036 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 15:46:15 crc kubenswrapper[4739]: I1201 15:46:15.821797 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp"] Dec 01 15:46:15 crc kubenswrapper[4739]: I1201 15:46:15.960615 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp\" (UID: \"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" Dec 01 15:46:15 crc kubenswrapper[4739]: I1201 15:46:15.960689 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbssg\" (UniqueName: \"kubernetes.io/projected/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-kube-api-access-bbssg\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp\" (UID: \"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" Dec 01 15:46:15 crc kubenswrapper[4739]: I1201 15:46:15.960727 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp\" (UID: \"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" Dec 01 15:46:16 crc kubenswrapper[4739]: I1201 15:46:16.062164 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp\" (UID: \"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" Dec 01 15:46:16 crc kubenswrapper[4739]: I1201 15:46:16.062294 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbssg\" (UniqueName: \"kubernetes.io/projected/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-kube-api-access-bbssg\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp\" (UID: \"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" Dec 01 15:46:16 crc kubenswrapper[4739]: I1201 15:46:16.062343 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp\" (UID: \"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" Dec 01 15:46:16 crc kubenswrapper[4739]: I1201 15:46:16.063167 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp\" (UID: \"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" Dec 01 15:46:16 crc kubenswrapper[4739]: I1201 15:46:16.063188 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp\" (UID: \"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" Dec 01 15:46:16 crc kubenswrapper[4739]: I1201 15:46:16.094644 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbssg\" (UniqueName: \"kubernetes.io/projected/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-kube-api-access-bbssg\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp\" (UID: \"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" Dec 01 15:46:16 crc kubenswrapper[4739]: I1201 15:46:16.184465 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" Dec 01 15:46:16 crc kubenswrapper[4739]: I1201 15:46:16.458096 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp"] Dec 01 15:46:17 crc kubenswrapper[4739]: I1201 15:46:17.210229 4739 generic.go:334] "Generic (PLEG): container finished" podID="9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9" containerID="43688e2b2cbcfda5489e1f38b5fa93699ff0384cef170fddf57f9684769598e4" exitCode=0 Dec 01 15:46:17 crc kubenswrapper[4739]: I1201 15:46:17.210370 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" event={"ID":"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9","Type":"ContainerDied","Data":"43688e2b2cbcfda5489e1f38b5fa93699ff0384cef170fddf57f9684769598e4"} Dec 01 15:46:17 crc kubenswrapper[4739]: I1201 15:46:17.210884 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" event={"ID":"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9","Type":"ContainerStarted","Data":"387a82d7c2c60531f4019b247a0c51cea2b94ba14739b18329b0adc879eb072f"} Dec 01 15:46:19 crc kubenswrapper[4739]: I1201 15:46:19.234286 4739 generic.go:334] "Generic (PLEG): container finished" podID="9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9" containerID="820a4cbc0b44d9001e973e4612288dcb71208302f505958d7c54175c7b4f6289" exitCode=0 Dec 01 15:46:19 crc kubenswrapper[4739]: I1201 15:46:19.234454 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" event={"ID":"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9","Type":"ContainerDied","Data":"820a4cbc0b44d9001e973e4612288dcb71208302f505958d7c54175c7b4f6289"} Dec 01 15:46:20 crc kubenswrapper[4739]: I1201 15:46:20.247748 4739 generic.go:334] "Generic (PLEG): container finished" podID="9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9" containerID="83c8e04a3021f3dc0885dcf1f987c081ecdf7eae27b387709c453ba021e57b9b" exitCode=0 Dec 01 15:46:20 crc kubenswrapper[4739]: I1201 15:46:20.247821 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" event={"ID":"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9","Type":"ContainerDied","Data":"83c8e04a3021f3dc0885dcf1f987c081ecdf7eae27b387709c453ba021e57b9b"} Dec 01 15:46:21 crc kubenswrapper[4739]: I1201 15:46:21.582683 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" Dec 01 15:46:21 crc kubenswrapper[4739]: I1201 15:46:21.739687 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbssg\" (UniqueName: \"kubernetes.io/projected/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-kube-api-access-bbssg\") pod \"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9\" (UID: \"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9\") " Dec 01 15:46:21 crc kubenswrapper[4739]: I1201 15:46:21.739879 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-util\") pod \"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9\" (UID: \"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9\") " Dec 01 15:46:21 crc kubenswrapper[4739]: I1201 15:46:21.739982 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-bundle\") pod \"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9\" (UID: \"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9\") " Dec 01 15:46:21 crc kubenswrapper[4739]: I1201 15:46:21.740936 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-bundle" (OuterVolumeSpecName: "bundle") pod "9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9" (UID: "9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:46:21 crc kubenswrapper[4739]: I1201 15:46:21.749648 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-kube-api-access-bbssg" (OuterVolumeSpecName: "kube-api-access-bbssg") pod "9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9" (UID: "9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9"). InnerVolumeSpecName "kube-api-access-bbssg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:46:21 crc kubenswrapper[4739]: I1201 15:46:21.768102 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-util" (OuterVolumeSpecName: "util") pod "9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9" (UID: "9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:46:21 crc kubenswrapper[4739]: I1201 15:46:21.841744 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbssg\" (UniqueName: \"kubernetes.io/projected/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-kube-api-access-bbssg\") on node \"crc\" DevicePath \"\"" Dec 01 15:46:21 crc kubenswrapper[4739]: I1201 15:46:21.841791 4739 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-util\") on node \"crc\" DevicePath \"\"" Dec 01 15:46:21 crc kubenswrapper[4739]: I1201 15:46:21.841810 4739 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:46:22 crc kubenswrapper[4739]: I1201 15:46:22.266283 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" event={"ID":"9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9","Type":"ContainerDied","Data":"387a82d7c2c60531f4019b247a0c51cea2b94ba14739b18329b0adc879eb072f"} Dec 01 15:46:22 crc kubenswrapper[4739]: I1201 15:46:22.266650 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="387a82d7c2c60531f4019b247a0c51cea2b94ba14739b18329b0adc879eb072f" Dec 01 15:46:22 crc kubenswrapper[4739]: I1201 15:46:22.266457 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp" Dec 01 15:46:24 crc kubenswrapper[4739]: I1201 15:46:24.540043 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-b9wt5"] Dec 01 15:46:24 crc kubenswrapper[4739]: E1201 15:46:24.540236 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9" containerName="pull" Dec 01 15:46:24 crc kubenswrapper[4739]: I1201 15:46:24.540248 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9" containerName="pull" Dec 01 15:46:24 crc kubenswrapper[4739]: E1201 15:46:24.540256 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9" containerName="extract" Dec 01 15:46:24 crc kubenswrapper[4739]: I1201 15:46:24.540263 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9" containerName="extract" Dec 01 15:46:24 crc kubenswrapper[4739]: E1201 15:46:24.540270 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9" containerName="util" Dec 01 15:46:24 crc kubenswrapper[4739]: I1201 15:46:24.540276 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9" containerName="util" Dec 01 15:46:24 crc kubenswrapper[4739]: I1201 15:46:24.540363 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9" containerName="extract" Dec 01 15:46:24 crc kubenswrapper[4739]: I1201 15:46:24.540724 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b9wt5" Dec 01 15:46:24 crc kubenswrapper[4739]: I1201 15:46:24.543378 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-72dbd" Dec 01 15:46:24 crc kubenswrapper[4739]: I1201 15:46:24.548360 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 01 15:46:24 crc kubenswrapper[4739]: I1201 15:46:24.551751 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 01 15:46:24 crc kubenswrapper[4739]: I1201 15:46:24.562581 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-b9wt5"] Dec 01 15:46:24 crc kubenswrapper[4739]: I1201 15:46:24.580221 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc7t8\" (UniqueName: \"kubernetes.io/projected/7edd99c1-023c-4505-b716-c87a7f5d3deb-kube-api-access-pc7t8\") pod \"nmstate-operator-5b5b58f5c8-b9wt5\" (UID: \"7edd99c1-023c-4505-b716-c87a7f5d3deb\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b9wt5" Dec 01 15:46:24 crc kubenswrapper[4739]: I1201 15:46:24.681645 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc7t8\" (UniqueName: \"kubernetes.io/projected/7edd99c1-023c-4505-b716-c87a7f5d3deb-kube-api-access-pc7t8\") pod \"nmstate-operator-5b5b58f5c8-b9wt5\" (UID: \"7edd99c1-023c-4505-b716-c87a7f5d3deb\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b9wt5" Dec 01 15:46:24 crc kubenswrapper[4739]: I1201 15:46:24.705209 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc7t8\" (UniqueName: \"kubernetes.io/projected/7edd99c1-023c-4505-b716-c87a7f5d3deb-kube-api-access-pc7t8\") pod \"nmstate-operator-5b5b58f5c8-b9wt5\" (UID: \"7edd99c1-023c-4505-b716-c87a7f5d3deb\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b9wt5" Dec 01 15:46:24 crc kubenswrapper[4739]: I1201 15:46:24.871621 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b9wt5" Dec 01 15:46:25 crc kubenswrapper[4739]: I1201 15:46:25.380587 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-b9wt5"] Dec 01 15:46:26 crc kubenswrapper[4739]: I1201 15:46:26.290663 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b9wt5" event={"ID":"7edd99c1-023c-4505-b716-c87a7f5d3deb","Type":"ContainerStarted","Data":"863debe5afce81657d60900646cd9f53a62238c321559669349ba4368a240551"} Dec 01 15:46:28 crc kubenswrapper[4739]: I1201 15:46:28.304488 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b9wt5" event={"ID":"7edd99c1-023c-4505-b716-c87a7f5d3deb","Type":"ContainerStarted","Data":"00683e0ddb5735858bdbc54908bf5d1d61ebfe9ad05abfb0b56311d4963e0247"} Dec 01 15:46:28 crc kubenswrapper[4739]: I1201 15:46:28.334135 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b9wt5" podStartSLOduration=2.03125345 podStartE2EDuration="4.334106307s" podCreationTimestamp="2025-12-01 15:46:24 +0000 UTC" firstStartedPulling="2025-12-01 15:46:25.388258195 +0000 UTC m=+687.214004319" lastFinishedPulling="2025-12-01 15:46:27.691111082 +0000 UTC m=+689.516857176" observedRunningTime="2025-12-01 15:46:28.327841838 +0000 UTC m=+690.153587942" watchObservedRunningTime="2025-12-01 15:46:28.334106307 +0000 UTC m=+690.159852441" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.302139 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-6fdfz"] Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.303615 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6fdfz" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.317623 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-p4mjq" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.320681 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-6fdfz"] Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.329540 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5"] Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.330411 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.336263 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.349331 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5"] Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.349874 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnks8\" (UniqueName: \"kubernetes.io/projected/aeaf3a4d-e661-4ed4-b54d-96b39645d838-kube-api-access-lnks8\") pod \"nmstate-webhook-5f6d4c5ccb-j2vz5\" (UID: \"aeaf3a4d-e661-4ed4-b54d-96b39645d838\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.349944 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/aeaf3a4d-e661-4ed4-b54d-96b39645d838-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-j2vz5\" (UID: \"aeaf3a4d-e661-4ed4-b54d-96b39645d838\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.349978 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nlzx\" (UniqueName: \"kubernetes.io/projected/d9a2f3d0-ea8e-435c-8e82-d1662c4cdeac-kube-api-access-5nlzx\") pod \"nmstate-metrics-7f946cbc9-6fdfz\" (UID: \"d9a2f3d0-ea8e-435c-8e82-d1662c4cdeac\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6fdfz" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.352406 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-cc6rd"] Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.353081 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.428388 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb"] Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.429023 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.432728 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.432846 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-p4667" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.432849 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.451385 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpfrj\" (UniqueName: \"kubernetes.io/projected/5889130c-3ff0-4d1d-b78c-8bce4ca14615-kube-api-access-cpfrj\") pod \"nmstate-console-plugin-7fbb5f6569-8fmbb\" (UID: \"5889130c-3ff0-4d1d-b78c-8bce4ca14615\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.451514 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/cdc17bb8-f21a-4de6-a6d7-87a69f18333e-ovs-socket\") pod \"nmstate-handler-cc6rd\" (UID: \"cdc17bb8-f21a-4de6-a6d7-87a69f18333e\") " pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.451554 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/aeaf3a4d-e661-4ed4-b54d-96b39645d838-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-j2vz5\" (UID: \"aeaf3a4d-e661-4ed4-b54d-96b39645d838\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.451587 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5889130c-3ff0-4d1d-b78c-8bce4ca14615-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-8fmbb\" (UID: \"5889130c-3ff0-4d1d-b78c-8bce4ca14615\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb" Dec 01 15:46:29 crc kubenswrapper[4739]: E1201 15:46:29.451750 4739 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 01 15:46:29 crc kubenswrapper[4739]: E1201 15:46:29.451830 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aeaf3a4d-e661-4ed4-b54d-96b39645d838-tls-key-pair podName:aeaf3a4d-e661-4ed4-b54d-96b39645d838 nodeName:}" failed. No retries permitted until 2025-12-01 15:46:29.951806245 +0000 UTC m=+691.777552339 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/aeaf3a4d-e661-4ed4-b54d-96b39645d838-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-j2vz5" (UID: "aeaf3a4d-e661-4ed4-b54d-96b39645d838") : secret "openshift-nmstate-webhook" not found Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.452065 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nlzx\" (UniqueName: \"kubernetes.io/projected/d9a2f3d0-ea8e-435c-8e82-d1662c4cdeac-kube-api-access-5nlzx\") pod \"nmstate-metrics-7f946cbc9-6fdfz\" (UID: \"d9a2f3d0-ea8e-435c-8e82-d1662c4cdeac\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6fdfz" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.452109 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/cdc17bb8-f21a-4de6-a6d7-87a69f18333e-dbus-socket\") pod \"nmstate-handler-cc6rd\" (UID: \"cdc17bb8-f21a-4de6-a6d7-87a69f18333e\") " pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.452188 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5889130c-3ff0-4d1d-b78c-8bce4ca14615-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-8fmbb\" (UID: \"5889130c-3ff0-4d1d-b78c-8bce4ca14615\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.452235 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/cdc17bb8-f21a-4de6-a6d7-87a69f18333e-nmstate-lock\") pod \"nmstate-handler-cc6rd\" (UID: \"cdc17bb8-f21a-4de6-a6d7-87a69f18333e\") " pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.452271 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnks8\" (UniqueName: \"kubernetes.io/projected/aeaf3a4d-e661-4ed4-b54d-96b39645d838-kube-api-access-lnks8\") pod \"nmstate-webhook-5f6d4c5ccb-j2vz5\" (UID: \"aeaf3a4d-e661-4ed4-b54d-96b39645d838\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.452291 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjltk\" (UniqueName: \"kubernetes.io/projected/cdc17bb8-f21a-4de6-a6d7-87a69f18333e-kube-api-access-zjltk\") pod \"nmstate-handler-cc6rd\" (UID: \"cdc17bb8-f21a-4de6-a6d7-87a69f18333e\") " pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.457109 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb"] Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.470589 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nlzx\" (UniqueName: \"kubernetes.io/projected/d9a2f3d0-ea8e-435c-8e82-d1662c4cdeac-kube-api-access-5nlzx\") pod \"nmstate-metrics-7f946cbc9-6fdfz\" (UID: \"d9a2f3d0-ea8e-435c-8e82-d1662c4cdeac\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6fdfz" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.473591 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnks8\" (UniqueName: \"kubernetes.io/projected/aeaf3a4d-e661-4ed4-b54d-96b39645d838-kube-api-access-lnks8\") pod \"nmstate-webhook-5f6d4c5ccb-j2vz5\" (UID: \"aeaf3a4d-e661-4ed4-b54d-96b39645d838\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.553619 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5889130c-3ff0-4d1d-b78c-8bce4ca14615-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-8fmbb\" (UID: \"5889130c-3ff0-4d1d-b78c-8bce4ca14615\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.553659 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/cdc17bb8-f21a-4de6-a6d7-87a69f18333e-nmstate-lock\") pod \"nmstate-handler-cc6rd\" (UID: \"cdc17bb8-f21a-4de6-a6d7-87a69f18333e\") " pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.553681 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjltk\" (UniqueName: \"kubernetes.io/projected/cdc17bb8-f21a-4de6-a6d7-87a69f18333e-kube-api-access-zjltk\") pod \"nmstate-handler-cc6rd\" (UID: \"cdc17bb8-f21a-4de6-a6d7-87a69f18333e\") " pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.553713 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpfrj\" (UniqueName: \"kubernetes.io/projected/5889130c-3ff0-4d1d-b78c-8bce4ca14615-kube-api-access-cpfrj\") pod \"nmstate-console-plugin-7fbb5f6569-8fmbb\" (UID: \"5889130c-3ff0-4d1d-b78c-8bce4ca14615\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.553732 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/cdc17bb8-f21a-4de6-a6d7-87a69f18333e-ovs-socket\") pod \"nmstate-handler-cc6rd\" (UID: \"cdc17bb8-f21a-4de6-a6d7-87a69f18333e\") " pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.553768 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5889130c-3ff0-4d1d-b78c-8bce4ca14615-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-8fmbb\" (UID: \"5889130c-3ff0-4d1d-b78c-8bce4ca14615\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.553796 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/cdc17bb8-f21a-4de6-a6d7-87a69f18333e-dbus-socket\") pod \"nmstate-handler-cc6rd\" (UID: \"cdc17bb8-f21a-4de6-a6d7-87a69f18333e\") " pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.554077 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/cdc17bb8-f21a-4de6-a6d7-87a69f18333e-dbus-socket\") pod \"nmstate-handler-cc6rd\" (UID: \"cdc17bb8-f21a-4de6-a6d7-87a69f18333e\") " pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.554119 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/cdc17bb8-f21a-4de6-a6d7-87a69f18333e-nmstate-lock\") pod \"nmstate-handler-cc6rd\" (UID: \"cdc17bb8-f21a-4de6-a6d7-87a69f18333e\") " pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.554507 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/cdc17bb8-f21a-4de6-a6d7-87a69f18333e-ovs-socket\") pod \"nmstate-handler-cc6rd\" (UID: \"cdc17bb8-f21a-4de6-a6d7-87a69f18333e\") " pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.555188 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5889130c-3ff0-4d1d-b78c-8bce4ca14615-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-8fmbb\" (UID: \"5889130c-3ff0-4d1d-b78c-8bce4ca14615\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.558732 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5889130c-3ff0-4d1d-b78c-8bce4ca14615-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-8fmbb\" (UID: \"5889130c-3ff0-4d1d-b78c-8bce4ca14615\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.580854 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpfrj\" (UniqueName: \"kubernetes.io/projected/5889130c-3ff0-4d1d-b78c-8bce4ca14615-kube-api-access-cpfrj\") pod \"nmstate-console-plugin-7fbb5f6569-8fmbb\" (UID: \"5889130c-3ff0-4d1d-b78c-8bce4ca14615\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.581028 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjltk\" (UniqueName: \"kubernetes.io/projected/cdc17bb8-f21a-4de6-a6d7-87a69f18333e-kube-api-access-zjltk\") pod \"nmstate-handler-cc6rd\" (UID: \"cdc17bb8-f21a-4de6-a6d7-87a69f18333e\") " pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.620523 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6fdfz" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.625266 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-64d7dddfc8-rg8xm"] Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.626016 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.639152 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-64d7dddfc8-rg8xm"] Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.655023 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/75505da9-0352-4b0a-9d50-e8e8aee4b905-oauth-serving-cert\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.655083 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/75505da9-0352-4b0a-9d50-e8e8aee4b905-console-config\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.655140 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75505da9-0352-4b0a-9d50-e8e8aee4b905-trusted-ca-bundle\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.655172 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/75505da9-0352-4b0a-9d50-e8e8aee4b905-service-ca\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.655198 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ksqc\" (UniqueName: \"kubernetes.io/projected/75505da9-0352-4b0a-9d50-e8e8aee4b905-kube-api-access-8ksqc\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.655221 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/75505da9-0352-4b0a-9d50-e8e8aee4b905-console-serving-cert\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.655241 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/75505da9-0352-4b0a-9d50-e8e8aee4b905-console-oauth-config\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.672154 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:29 crc kubenswrapper[4739]: W1201 15:46:29.686913 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcdc17bb8_f21a_4de6_a6d7_87a69f18333e.slice/crio-6ba0975ada730ed60de5bd5aa96add112bd2fa36b847833ba1fb2df3a77b0d3b WatchSource:0}: Error finding container 6ba0975ada730ed60de5bd5aa96add112bd2fa36b847833ba1fb2df3a77b0d3b: Status 404 returned error can't find the container with id 6ba0975ada730ed60de5bd5aa96add112bd2fa36b847833ba1fb2df3a77b0d3b Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.749107 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.758487 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/75505da9-0352-4b0a-9d50-e8e8aee4b905-oauth-serving-cert\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.758538 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/75505da9-0352-4b0a-9d50-e8e8aee4b905-console-config\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.758577 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75505da9-0352-4b0a-9d50-e8e8aee4b905-trusted-ca-bundle\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.758599 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/75505da9-0352-4b0a-9d50-e8e8aee4b905-service-ca\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.758619 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ksqc\" (UniqueName: \"kubernetes.io/projected/75505da9-0352-4b0a-9d50-e8e8aee4b905-kube-api-access-8ksqc\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.758639 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/75505da9-0352-4b0a-9d50-e8e8aee4b905-console-serving-cert\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.758654 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/75505da9-0352-4b0a-9d50-e8e8aee4b905-console-oauth-config\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.759817 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75505da9-0352-4b0a-9d50-e8e8aee4b905-trusted-ca-bundle\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.760388 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/75505da9-0352-4b0a-9d50-e8e8aee4b905-oauth-serving-cert\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.760950 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/75505da9-0352-4b0a-9d50-e8e8aee4b905-console-config\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.762584 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/75505da9-0352-4b0a-9d50-e8e8aee4b905-service-ca\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.779576 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/75505da9-0352-4b0a-9d50-e8e8aee4b905-console-serving-cert\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.780739 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/75505da9-0352-4b0a-9d50-e8e8aee4b905-console-oauth-config\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.795388 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ksqc\" (UniqueName: \"kubernetes.io/projected/75505da9-0352-4b0a-9d50-e8e8aee4b905-kube-api-access-8ksqc\") pod \"console-64d7dddfc8-rg8xm\" (UID: \"75505da9-0352-4b0a-9d50-e8e8aee4b905\") " pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.939725 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb"] Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.949689 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.962328 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/aeaf3a4d-e661-4ed4-b54d-96b39645d838-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-j2vz5\" (UID: \"aeaf3a4d-e661-4ed4-b54d-96b39645d838\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5" Dec 01 15:46:29 crc kubenswrapper[4739]: I1201 15:46:29.965573 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/aeaf3a4d-e661-4ed4-b54d-96b39645d838-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-j2vz5\" (UID: \"aeaf3a4d-e661-4ed4-b54d-96b39645d838\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5" Dec 01 15:46:30 crc kubenswrapper[4739]: I1201 15:46:30.096636 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-6fdfz"] Dec 01 15:46:30 crc kubenswrapper[4739]: W1201 15:46:30.114015 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9a2f3d0_ea8e_435c_8e82_d1662c4cdeac.slice/crio-83df9dc0b34ab09f595ca735accfe4692936a1db1918e9f06910c16466797221 WatchSource:0}: Error finding container 83df9dc0b34ab09f595ca735accfe4692936a1db1918e9f06910c16466797221: Status 404 returned error can't find the container with id 83df9dc0b34ab09f595ca735accfe4692936a1db1918e9f06910c16466797221 Dec 01 15:46:30 crc kubenswrapper[4739]: I1201 15:46:30.246978 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5" Dec 01 15:46:30 crc kubenswrapper[4739]: I1201 15:46:30.316040 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb" event={"ID":"5889130c-3ff0-4d1d-b78c-8bce4ca14615","Type":"ContainerStarted","Data":"3b48b0e9fccc124ca34b1c4c4f20a70370ebd19cdbdc3b669a7f41a03ad3213d"} Dec 01 15:46:30 crc kubenswrapper[4739]: I1201 15:46:30.317117 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-cc6rd" event={"ID":"cdc17bb8-f21a-4de6-a6d7-87a69f18333e","Type":"ContainerStarted","Data":"6ba0975ada730ed60de5bd5aa96add112bd2fa36b847833ba1fb2df3a77b0d3b"} Dec 01 15:46:30 crc kubenswrapper[4739]: I1201 15:46:30.317890 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6fdfz" event={"ID":"d9a2f3d0-ea8e-435c-8e82-d1662c4cdeac","Type":"ContainerStarted","Data":"83df9dc0b34ab09f595ca735accfe4692936a1db1918e9f06910c16466797221"} Dec 01 15:46:30 crc kubenswrapper[4739]: I1201 15:46:30.332148 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-64d7dddfc8-rg8xm"] Dec 01 15:46:30 crc kubenswrapper[4739]: W1201 15:46:30.339995 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75505da9_0352_4b0a_9d50_e8e8aee4b905.slice/crio-da8cbfcbe58803b3f2590ef7f0204e292c62e2f52a10f947e096023033e3d731 WatchSource:0}: Error finding container da8cbfcbe58803b3f2590ef7f0204e292c62e2f52a10f947e096023033e3d731: Status 404 returned error can't find the container with id da8cbfcbe58803b3f2590ef7f0204e292c62e2f52a10f947e096023033e3d731 Dec 01 15:46:30 crc kubenswrapper[4739]: I1201 15:46:30.455550 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5"] Dec 01 15:46:30 crc kubenswrapper[4739]: W1201 15:46:30.463083 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaeaf3a4d_e661_4ed4_b54d_96b39645d838.slice/crio-02deda87fcde4a78f99e167428b6114c9f74038e5575067382c5773943690bb9 WatchSource:0}: Error finding container 02deda87fcde4a78f99e167428b6114c9f74038e5575067382c5773943690bb9: Status 404 returned error can't find the container with id 02deda87fcde4a78f99e167428b6114c9f74038e5575067382c5773943690bb9 Dec 01 15:46:31 crc kubenswrapper[4739]: I1201 15:46:31.328353 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-64d7dddfc8-rg8xm" event={"ID":"75505da9-0352-4b0a-9d50-e8e8aee4b905","Type":"ContainerStarted","Data":"5cda4a7fadf7d6408479c1d76077840f25c61e41717df181107d92c8400386f7"} Dec 01 15:46:31 crc kubenswrapper[4739]: I1201 15:46:31.328851 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-64d7dddfc8-rg8xm" event={"ID":"75505da9-0352-4b0a-9d50-e8e8aee4b905","Type":"ContainerStarted","Data":"da8cbfcbe58803b3f2590ef7f0204e292c62e2f52a10f947e096023033e3d731"} Dec 01 15:46:31 crc kubenswrapper[4739]: I1201 15:46:31.332545 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5" event={"ID":"aeaf3a4d-e661-4ed4-b54d-96b39645d838","Type":"ContainerStarted","Data":"02deda87fcde4a78f99e167428b6114c9f74038e5575067382c5773943690bb9"} Dec 01 15:46:31 crc kubenswrapper[4739]: I1201 15:46:31.360917 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-64d7dddfc8-rg8xm" podStartSLOduration=2.360884218 podStartE2EDuration="2.360884218s" podCreationTimestamp="2025-12-01 15:46:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:46:31.356751246 +0000 UTC m=+693.182497360" watchObservedRunningTime="2025-12-01 15:46:31.360884218 +0000 UTC m=+693.186630352" Dec 01 15:46:33 crc kubenswrapper[4739]: I1201 15:46:33.349644 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb" event={"ID":"5889130c-3ff0-4d1d-b78c-8bce4ca14615","Type":"ContainerStarted","Data":"92e223a72b51b131ca8bafc01955a6956211e31a006a8e5b289c16dac3a2d113"} Dec 01 15:46:33 crc kubenswrapper[4739]: I1201 15:46:33.351609 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-cc6rd" event={"ID":"cdc17bb8-f21a-4de6-a6d7-87a69f18333e","Type":"ContainerStarted","Data":"ae013db8d5d3f9d5baaa7ac9d18a9c6631d5e2c341726d7b687b51252954e682"} Dec 01 15:46:33 crc kubenswrapper[4739]: I1201 15:46:33.351714 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:33 crc kubenswrapper[4739]: I1201 15:46:33.353800 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6fdfz" event={"ID":"d9a2f3d0-ea8e-435c-8e82-d1662c4cdeac","Type":"ContainerStarted","Data":"162c48eba86e7eb45ae1b7a5b8cc9d54dc0e7531435d3697b625dace9a720c20"} Dec 01 15:46:33 crc kubenswrapper[4739]: I1201 15:46:33.355330 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5" event={"ID":"aeaf3a4d-e661-4ed4-b54d-96b39645d838","Type":"ContainerStarted","Data":"0fcba98d400eb87fbf1a2808767967f97f19e9c1b62a2f85104df6bff20db31b"} Dec 01 15:46:33 crc kubenswrapper[4739]: I1201 15:46:33.355505 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5" Dec 01 15:46:33 crc kubenswrapper[4739]: I1201 15:46:33.374156 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8fmbb" podStartSLOduration=1.412828071 podStartE2EDuration="4.37398096s" podCreationTimestamp="2025-12-01 15:46:29 +0000 UTC" firstStartedPulling="2025-12-01 15:46:29.944349745 +0000 UTC m=+691.770095839" lastFinishedPulling="2025-12-01 15:46:32.905502634 +0000 UTC m=+694.731248728" observedRunningTime="2025-12-01 15:46:33.368593649 +0000 UTC m=+695.194339783" watchObservedRunningTime="2025-12-01 15:46:33.37398096 +0000 UTC m=+695.199727094" Dec 01 15:46:33 crc kubenswrapper[4739]: I1201 15:46:33.419358 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5" podStartSLOduration=1.971077607 podStartE2EDuration="4.419331339s" podCreationTimestamp="2025-12-01 15:46:29 +0000 UTC" firstStartedPulling="2025-12-01 15:46:30.465166443 +0000 UTC m=+692.290912537" lastFinishedPulling="2025-12-01 15:46:32.913420135 +0000 UTC m=+694.739166269" observedRunningTime="2025-12-01 15:46:33.392137076 +0000 UTC m=+695.217883220" watchObservedRunningTime="2025-12-01 15:46:33.419331339 +0000 UTC m=+695.245077473" Dec 01 15:46:33 crc kubenswrapper[4739]: I1201 15:46:33.420211 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-cc6rd" podStartSLOduration=1.196860259 podStartE2EDuration="4.420199577s" podCreationTimestamp="2025-12-01 15:46:29 +0000 UTC" firstStartedPulling="2025-12-01 15:46:29.688367823 +0000 UTC m=+691.514113917" lastFinishedPulling="2025-12-01 15:46:32.911707131 +0000 UTC m=+694.737453235" observedRunningTime="2025-12-01 15:46:33.412312807 +0000 UTC m=+695.238058961" watchObservedRunningTime="2025-12-01 15:46:33.420199577 +0000 UTC m=+695.245945681" Dec 01 15:46:36 crc kubenswrapper[4739]: I1201 15:46:36.383448 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6fdfz" event={"ID":"d9a2f3d0-ea8e-435c-8e82-d1662c4cdeac","Type":"ContainerStarted","Data":"b5c1a50d4fdc8136dcc6c9433e14114b37514b0fcd6730712d39a7bc61f886f0"} Dec 01 15:46:36 crc kubenswrapper[4739]: I1201 15:46:36.417749 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-6fdfz" podStartSLOduration=1.928681571 podStartE2EDuration="7.417710848s" podCreationTimestamp="2025-12-01 15:46:29 +0000 UTC" firstStartedPulling="2025-12-01 15:46:30.115460755 +0000 UTC m=+691.941206849" lastFinishedPulling="2025-12-01 15:46:35.604490002 +0000 UTC m=+697.430236126" observedRunningTime="2025-12-01 15:46:36.409347133 +0000 UTC m=+698.235093287" watchObservedRunningTime="2025-12-01 15:46:36.417710848 +0000 UTC m=+698.243456982" Dec 01 15:46:39 crc kubenswrapper[4739]: I1201 15:46:39.622527 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:46:39 crc kubenswrapper[4739]: I1201 15:46:39.622989 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:46:39 crc kubenswrapper[4739]: I1201 15:46:39.710775 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-cc6rd" Dec 01 15:46:39 crc kubenswrapper[4739]: I1201 15:46:39.950149 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:39 crc kubenswrapper[4739]: I1201 15:46:39.950677 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:39 crc kubenswrapper[4739]: I1201 15:46:39.957784 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:40 crc kubenswrapper[4739]: I1201 15:46:40.421256 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-64d7dddfc8-rg8xm" Dec 01 15:46:40 crc kubenswrapper[4739]: I1201 15:46:40.506904 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5dxtl"] Dec 01 15:46:50 crc kubenswrapper[4739]: I1201 15:46:50.378784 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j2vz5" Dec 01 15:47:05 crc kubenswrapper[4739]: I1201 15:47:05.573239 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-5dxtl" podUID="64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" containerName="console" containerID="cri-o://051064c8ca5edbb931f1b116e15d032114793f2e260c16fec8b47c7fe3b3fa3c" gracePeriod=15 Dec 01 15:47:05 crc kubenswrapper[4739]: I1201 15:47:05.709952 4739 patch_prober.go:28] interesting pod/console-f9d7485db-5dxtl container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Dec 01 15:47:05 crc kubenswrapper[4739]: I1201 15:47:05.710037 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-f9d7485db-5dxtl" podUID="64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" containerName="console" probeResult="failure" output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" Dec 01 15:47:05 crc kubenswrapper[4739]: I1201 15:47:05.924061 4739 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.323660 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s"] Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.325173 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.331833 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s"] Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.332662 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.503753 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5dxtl_64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3/console/0.log" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.503825 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.516341 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7beb8fab-f0fc-4d48-8688-ae02488b2e51-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s\" (UID: \"7beb8fab-f0fc-4d48-8688-ae02488b2e51\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.516466 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fx5d4\" (UniqueName: \"kubernetes.io/projected/7beb8fab-f0fc-4d48-8688-ae02488b2e51-kube-api-access-fx5d4\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s\" (UID: \"7beb8fab-f0fc-4d48-8688-ae02488b2e51\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.516495 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7beb8fab-f0fc-4d48-8688-ae02488b2e51-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s\" (UID: \"7beb8fab-f0fc-4d48-8688-ae02488b2e51\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.583991 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5dxtl_64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3/console/0.log" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.584062 4739 generic.go:334] "Generic (PLEG): container finished" podID="64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" containerID="051064c8ca5edbb931f1b116e15d032114793f2e260c16fec8b47c7fe3b3fa3c" exitCode=2 Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.584103 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5dxtl" event={"ID":"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3","Type":"ContainerDied","Data":"051064c8ca5edbb931f1b116e15d032114793f2e260c16fec8b47c7fe3b3fa3c"} Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.584123 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5dxtl" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.584165 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5dxtl" event={"ID":"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3","Type":"ContainerDied","Data":"429370f7b95781a4e5d1080dd48ade1fd9afe8e4b4a8a4f40c7144d2cd052726"} Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.584195 4739 scope.go:117] "RemoveContainer" containerID="051064c8ca5edbb931f1b116e15d032114793f2e260c16fec8b47c7fe3b3fa3c" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.611623 4739 scope.go:117] "RemoveContainer" containerID="051064c8ca5edbb931f1b116e15d032114793f2e260c16fec8b47c7fe3b3fa3c" Dec 01 15:47:06 crc kubenswrapper[4739]: E1201 15:47:06.612207 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"051064c8ca5edbb931f1b116e15d032114793f2e260c16fec8b47c7fe3b3fa3c\": container with ID starting with 051064c8ca5edbb931f1b116e15d032114793f2e260c16fec8b47c7fe3b3fa3c not found: ID does not exist" containerID="051064c8ca5edbb931f1b116e15d032114793f2e260c16fec8b47c7fe3b3fa3c" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.612294 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"051064c8ca5edbb931f1b116e15d032114793f2e260c16fec8b47c7fe3b3fa3c"} err="failed to get container status \"051064c8ca5edbb931f1b116e15d032114793f2e260c16fec8b47c7fe3b3fa3c\": rpc error: code = NotFound desc = could not find container \"051064c8ca5edbb931f1b116e15d032114793f2e260c16fec8b47c7fe3b3fa3c\": container with ID starting with 051064c8ca5edbb931f1b116e15d032114793f2e260c16fec8b47c7fe3b3fa3c not found: ID does not exist" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.617790 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-serving-cert\") pod \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.617907 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-trusted-ca-bundle\") pod \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.618029 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-config\") pod \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.618124 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-oauth-serving-cert\") pod \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.618206 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-service-ca\") pod \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.618342 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-oauth-config\") pod \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.618483 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtmhz\" (UniqueName: \"kubernetes.io/projected/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-kube-api-access-wtmhz\") pod \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\" (UID: \"64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3\") " Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.618989 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" (UID: "64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.619006 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-config" (OuterVolumeSpecName: "console-config") pod "64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" (UID: "64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.619172 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" (UID: "64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.619411 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fx5d4\" (UniqueName: \"kubernetes.io/projected/7beb8fab-f0fc-4d48-8688-ae02488b2e51-kube-api-access-fx5d4\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s\" (UID: \"7beb8fab-f0fc-4d48-8688-ae02488b2e51\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.619556 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7beb8fab-f0fc-4d48-8688-ae02488b2e51-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s\" (UID: \"7beb8fab-f0fc-4d48-8688-ae02488b2e51\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.619681 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7beb8fab-f0fc-4d48-8688-ae02488b2e51-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s\" (UID: \"7beb8fab-f0fc-4d48-8688-ae02488b2e51\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.619804 4739 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.619881 4739 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.619950 4739 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.620244 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-service-ca" (OuterVolumeSpecName: "service-ca") pod "64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" (UID: "64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.620909 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7beb8fab-f0fc-4d48-8688-ae02488b2e51-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s\" (UID: \"7beb8fab-f0fc-4d48-8688-ae02488b2e51\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.621084 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7beb8fab-f0fc-4d48-8688-ae02488b2e51-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s\" (UID: \"7beb8fab-f0fc-4d48-8688-ae02488b2e51\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.628022 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-kube-api-access-wtmhz" (OuterVolumeSpecName: "kube-api-access-wtmhz") pod "64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" (UID: "64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3"). InnerVolumeSpecName "kube-api-access-wtmhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.628357 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" (UID: "64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.628766 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" (UID: "64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.642564 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fx5d4\" (UniqueName: \"kubernetes.io/projected/7beb8fab-f0fc-4d48-8688-ae02488b2e51-kube-api-access-fx5d4\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s\" (UID: \"7beb8fab-f0fc-4d48-8688-ae02488b2e51\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.643165 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.721236 4739 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.721326 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtmhz\" (UniqueName: \"kubernetes.io/projected/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-kube-api-access-wtmhz\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.721383 4739 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.721405 4739 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.889981 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s"] Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.948030 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5dxtl"] Dec 01 15:47:06 crc kubenswrapper[4739]: I1201 15:47:06.952591 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-5dxtl"] Dec 01 15:47:07 crc kubenswrapper[4739]: I1201 15:47:07.593902 4739 generic.go:334] "Generic (PLEG): container finished" podID="7beb8fab-f0fc-4d48-8688-ae02488b2e51" containerID="56e48aa0f7d2ed3726e5fc5edeeb61a74177103d2d56538e915f5e08d82f67fa" exitCode=0 Dec 01 15:47:07 crc kubenswrapper[4739]: I1201 15:47:07.593981 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" event={"ID":"7beb8fab-f0fc-4d48-8688-ae02488b2e51","Type":"ContainerDied","Data":"56e48aa0f7d2ed3726e5fc5edeeb61a74177103d2d56538e915f5e08d82f67fa"} Dec 01 15:47:07 crc kubenswrapper[4739]: I1201 15:47:07.594488 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" event={"ID":"7beb8fab-f0fc-4d48-8688-ae02488b2e51","Type":"ContainerStarted","Data":"31b6ee0b60f68d86ed9c7651c64ac9a9d49932d8936c8d857c1999fa40c39ed9"} Dec 01 15:47:08 crc kubenswrapper[4739]: I1201 15:47:08.488645 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" path="/var/lib/kubelet/pods/64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3/volumes" Dec 01 15:47:08 crc kubenswrapper[4739]: I1201 15:47:08.644528 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bfhc7"] Dec 01 15:47:08 crc kubenswrapper[4739]: E1201 15:47:08.644941 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" containerName="console" Dec 01 15:47:08 crc kubenswrapper[4739]: I1201 15:47:08.644966 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" containerName="console" Dec 01 15:47:08 crc kubenswrapper[4739]: I1201 15:47:08.645294 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="64e2d6dd-8dfe-4a78-a58f-8ce88cae23d3" containerName="console" Dec 01 15:47:08 crc kubenswrapper[4739]: I1201 15:47:08.649335 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:08 crc kubenswrapper[4739]: I1201 15:47:08.657032 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bfhc7"] Dec 01 15:47:08 crc kubenswrapper[4739]: I1201 15:47:08.749624 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kthdz\" (UniqueName: \"kubernetes.io/projected/c163b957-f79f-4bef-96d5-1a6d4353a849-kube-api-access-kthdz\") pod \"redhat-operators-bfhc7\" (UID: \"c163b957-f79f-4bef-96d5-1a6d4353a849\") " pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:08 crc kubenswrapper[4739]: I1201 15:47:08.750118 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c163b957-f79f-4bef-96d5-1a6d4353a849-utilities\") pod \"redhat-operators-bfhc7\" (UID: \"c163b957-f79f-4bef-96d5-1a6d4353a849\") " pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:08 crc kubenswrapper[4739]: I1201 15:47:08.750160 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c163b957-f79f-4bef-96d5-1a6d4353a849-catalog-content\") pod \"redhat-operators-bfhc7\" (UID: \"c163b957-f79f-4bef-96d5-1a6d4353a849\") " pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:08 crc kubenswrapper[4739]: I1201 15:47:08.851692 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c163b957-f79f-4bef-96d5-1a6d4353a849-utilities\") pod \"redhat-operators-bfhc7\" (UID: \"c163b957-f79f-4bef-96d5-1a6d4353a849\") " pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:08 crc kubenswrapper[4739]: I1201 15:47:08.851767 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c163b957-f79f-4bef-96d5-1a6d4353a849-catalog-content\") pod \"redhat-operators-bfhc7\" (UID: \"c163b957-f79f-4bef-96d5-1a6d4353a849\") " pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:08 crc kubenswrapper[4739]: I1201 15:47:08.851836 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kthdz\" (UniqueName: \"kubernetes.io/projected/c163b957-f79f-4bef-96d5-1a6d4353a849-kube-api-access-kthdz\") pod \"redhat-operators-bfhc7\" (UID: \"c163b957-f79f-4bef-96d5-1a6d4353a849\") " pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:08 crc kubenswrapper[4739]: I1201 15:47:08.852489 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c163b957-f79f-4bef-96d5-1a6d4353a849-utilities\") pod \"redhat-operators-bfhc7\" (UID: \"c163b957-f79f-4bef-96d5-1a6d4353a849\") " pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:08 crc kubenswrapper[4739]: I1201 15:47:08.852574 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c163b957-f79f-4bef-96d5-1a6d4353a849-catalog-content\") pod \"redhat-operators-bfhc7\" (UID: \"c163b957-f79f-4bef-96d5-1a6d4353a849\") " pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:08 crc kubenswrapper[4739]: I1201 15:47:08.879016 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kthdz\" (UniqueName: \"kubernetes.io/projected/c163b957-f79f-4bef-96d5-1a6d4353a849-kube-api-access-kthdz\") pod \"redhat-operators-bfhc7\" (UID: \"c163b957-f79f-4bef-96d5-1a6d4353a849\") " pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:09 crc kubenswrapper[4739]: I1201 15:47:09.016482 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:09 crc kubenswrapper[4739]: I1201 15:47:09.301307 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bfhc7"] Dec 01 15:47:09 crc kubenswrapper[4739]: W1201 15:47:09.310403 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc163b957_f79f_4bef_96d5_1a6d4353a849.slice/crio-93969a3070188f89d487d95ddb904ccee8575e4e8bf0997b88b38dd368ea0eb8 WatchSource:0}: Error finding container 93969a3070188f89d487d95ddb904ccee8575e4e8bf0997b88b38dd368ea0eb8: Status 404 returned error can't find the container with id 93969a3070188f89d487d95ddb904ccee8575e4e8bf0997b88b38dd368ea0eb8 Dec 01 15:47:09 crc kubenswrapper[4739]: I1201 15:47:09.612068 4739 generic.go:334] "Generic (PLEG): container finished" podID="c163b957-f79f-4bef-96d5-1a6d4353a849" containerID="f203d52aac5e9a32dfe247f2f67905e162c7d2fadda454a8e0b23e8f756f2270" exitCode=0 Dec 01 15:47:09 crc kubenswrapper[4739]: I1201 15:47:09.612143 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfhc7" event={"ID":"c163b957-f79f-4bef-96d5-1a6d4353a849","Type":"ContainerDied","Data":"f203d52aac5e9a32dfe247f2f67905e162c7d2fadda454a8e0b23e8f756f2270"} Dec 01 15:47:09 crc kubenswrapper[4739]: I1201 15:47:09.612172 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfhc7" event={"ID":"c163b957-f79f-4bef-96d5-1a6d4353a849","Type":"ContainerStarted","Data":"93969a3070188f89d487d95ddb904ccee8575e4e8bf0997b88b38dd368ea0eb8"} Dec 01 15:47:09 crc kubenswrapper[4739]: I1201 15:47:09.614217 4739 generic.go:334] "Generic (PLEG): container finished" podID="7beb8fab-f0fc-4d48-8688-ae02488b2e51" containerID="d34d301f18df7c53e4eb6025bdcf200502df9e94078b52468cf0cbc624f8dc1a" exitCode=0 Dec 01 15:47:09 crc kubenswrapper[4739]: I1201 15:47:09.614262 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" event={"ID":"7beb8fab-f0fc-4d48-8688-ae02488b2e51","Type":"ContainerDied","Data":"d34d301f18df7c53e4eb6025bdcf200502df9e94078b52468cf0cbc624f8dc1a"} Dec 01 15:47:09 crc kubenswrapper[4739]: I1201 15:47:09.622001 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:47:09 crc kubenswrapper[4739]: I1201 15:47:09.622049 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:47:10 crc kubenswrapper[4739]: I1201 15:47:10.629143 4739 generic.go:334] "Generic (PLEG): container finished" podID="7beb8fab-f0fc-4d48-8688-ae02488b2e51" containerID="d64b2eaeab8917cede1e6264980bd94356c3cf4f418dbae44bf27c028b2236b4" exitCode=0 Dec 01 15:47:10 crc kubenswrapper[4739]: I1201 15:47:10.629277 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" event={"ID":"7beb8fab-f0fc-4d48-8688-ae02488b2e51","Type":"ContainerDied","Data":"d64b2eaeab8917cede1e6264980bd94356c3cf4f418dbae44bf27c028b2236b4"} Dec 01 15:47:10 crc kubenswrapper[4739]: I1201 15:47:10.633587 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfhc7" event={"ID":"c163b957-f79f-4bef-96d5-1a6d4353a849","Type":"ContainerStarted","Data":"94d43bb96978f38ebcc6bffaa0e0f7f2f9a8cca4c76b0ca180feb5c3ea0404ee"} Dec 01 15:47:11 crc kubenswrapper[4739]: I1201 15:47:11.642922 4739 generic.go:334] "Generic (PLEG): container finished" podID="c163b957-f79f-4bef-96d5-1a6d4353a849" containerID="94d43bb96978f38ebcc6bffaa0e0f7f2f9a8cca4c76b0ca180feb5c3ea0404ee" exitCode=0 Dec 01 15:47:11 crc kubenswrapper[4739]: I1201 15:47:11.643273 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfhc7" event={"ID":"c163b957-f79f-4bef-96d5-1a6d4353a849","Type":"ContainerDied","Data":"94d43bb96978f38ebcc6bffaa0e0f7f2f9a8cca4c76b0ca180feb5c3ea0404ee"} Dec 01 15:47:11 crc kubenswrapper[4739]: I1201 15:47:11.984097 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" Dec 01 15:47:12 crc kubenswrapper[4739]: I1201 15:47:12.100700 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fx5d4\" (UniqueName: \"kubernetes.io/projected/7beb8fab-f0fc-4d48-8688-ae02488b2e51-kube-api-access-fx5d4\") pod \"7beb8fab-f0fc-4d48-8688-ae02488b2e51\" (UID: \"7beb8fab-f0fc-4d48-8688-ae02488b2e51\") " Dec 01 15:47:12 crc kubenswrapper[4739]: I1201 15:47:12.100809 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7beb8fab-f0fc-4d48-8688-ae02488b2e51-util\") pod \"7beb8fab-f0fc-4d48-8688-ae02488b2e51\" (UID: \"7beb8fab-f0fc-4d48-8688-ae02488b2e51\") " Dec 01 15:47:12 crc kubenswrapper[4739]: I1201 15:47:12.100983 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7beb8fab-f0fc-4d48-8688-ae02488b2e51-bundle\") pod \"7beb8fab-f0fc-4d48-8688-ae02488b2e51\" (UID: \"7beb8fab-f0fc-4d48-8688-ae02488b2e51\") " Dec 01 15:47:12 crc kubenswrapper[4739]: I1201 15:47:12.102761 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7beb8fab-f0fc-4d48-8688-ae02488b2e51-bundle" (OuterVolumeSpecName: "bundle") pod "7beb8fab-f0fc-4d48-8688-ae02488b2e51" (UID: "7beb8fab-f0fc-4d48-8688-ae02488b2e51"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:47:12 crc kubenswrapper[4739]: I1201 15:47:12.109786 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7beb8fab-f0fc-4d48-8688-ae02488b2e51-kube-api-access-fx5d4" (OuterVolumeSpecName: "kube-api-access-fx5d4") pod "7beb8fab-f0fc-4d48-8688-ae02488b2e51" (UID: "7beb8fab-f0fc-4d48-8688-ae02488b2e51"). InnerVolumeSpecName "kube-api-access-fx5d4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:47:12 crc kubenswrapper[4739]: I1201 15:47:12.122331 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7beb8fab-f0fc-4d48-8688-ae02488b2e51-util" (OuterVolumeSpecName: "util") pod "7beb8fab-f0fc-4d48-8688-ae02488b2e51" (UID: "7beb8fab-f0fc-4d48-8688-ae02488b2e51"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:47:12 crc kubenswrapper[4739]: I1201 15:47:12.202630 4739 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7beb8fab-f0fc-4d48-8688-ae02488b2e51-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:12 crc kubenswrapper[4739]: I1201 15:47:12.202665 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fx5d4\" (UniqueName: \"kubernetes.io/projected/7beb8fab-f0fc-4d48-8688-ae02488b2e51-kube-api-access-fx5d4\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:12 crc kubenswrapper[4739]: I1201 15:47:12.202681 4739 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7beb8fab-f0fc-4d48-8688-ae02488b2e51-util\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:12 crc kubenswrapper[4739]: I1201 15:47:12.655107 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfhc7" event={"ID":"c163b957-f79f-4bef-96d5-1a6d4353a849","Type":"ContainerStarted","Data":"fa32fb8c0dd993efe19ff16539b3817d2b405e6299e01cc8601822880e27a870"} Dec 01 15:47:12 crc kubenswrapper[4739]: I1201 15:47:12.659239 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" event={"ID":"7beb8fab-f0fc-4d48-8688-ae02488b2e51","Type":"ContainerDied","Data":"31b6ee0b60f68d86ed9c7651c64ac9a9d49932d8936c8d857c1999fa40c39ed9"} Dec 01 15:47:12 crc kubenswrapper[4739]: I1201 15:47:12.659297 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31b6ee0b60f68d86ed9c7651c64ac9a9d49932d8936c8d857c1999fa40c39ed9" Dec 01 15:47:12 crc kubenswrapper[4739]: I1201 15:47:12.659342 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s" Dec 01 15:47:12 crc kubenswrapper[4739]: I1201 15:47:12.684506 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bfhc7" podStartSLOduration=2.221201117 podStartE2EDuration="4.684489859s" podCreationTimestamp="2025-12-01 15:47:08 +0000 UTC" firstStartedPulling="2025-12-01 15:47:09.61391921 +0000 UTC m=+731.439665304" lastFinishedPulling="2025-12-01 15:47:12.077207942 +0000 UTC m=+733.902954046" observedRunningTime="2025-12-01 15:47:12.683619042 +0000 UTC m=+734.509365216" watchObservedRunningTime="2025-12-01 15:47:12.684489859 +0000 UTC m=+734.510235963" Dec 01 15:47:19 crc kubenswrapper[4739]: I1201 15:47:19.017315 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:19 crc kubenswrapper[4739]: I1201 15:47:19.017892 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:19 crc kubenswrapper[4739]: I1201 15:47:19.095921 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:19 crc kubenswrapper[4739]: I1201 15:47:19.785225 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:20 crc kubenswrapper[4739]: I1201 15:47:20.423906 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bfhc7"] Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.665331 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv"] Dec 01 15:47:21 crc kubenswrapper[4739]: E1201 15:47:21.665975 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7beb8fab-f0fc-4d48-8688-ae02488b2e51" containerName="pull" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.665994 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7beb8fab-f0fc-4d48-8688-ae02488b2e51" containerName="pull" Dec 01 15:47:21 crc kubenswrapper[4739]: E1201 15:47:21.666027 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7beb8fab-f0fc-4d48-8688-ae02488b2e51" containerName="extract" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.666039 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7beb8fab-f0fc-4d48-8688-ae02488b2e51" containerName="extract" Dec 01 15:47:21 crc kubenswrapper[4739]: E1201 15:47:21.666061 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7beb8fab-f0fc-4d48-8688-ae02488b2e51" containerName="util" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.666074 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7beb8fab-f0fc-4d48-8688-ae02488b2e51" containerName="util" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.666261 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="7beb8fab-f0fc-4d48-8688-ae02488b2e51" containerName="extract" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.666836 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.669999 4739 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-lww2b" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.670017 4739 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.670266 4739 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.671088 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.671091 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.682388 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv"] Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.735465 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bfhc7" podUID="c163b957-f79f-4bef-96d5-1a6d4353a849" containerName="registry-server" containerID="cri-o://fa32fb8c0dd993efe19ff16539b3817d2b405e6299e01cc8601822880e27a870" gracePeriod=2 Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.839601 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c577e6ac-c778-4859-8b39-30aeee31c0ec-webhook-cert\") pod \"metallb-operator-controller-manager-697d66d89d-jhzwv\" (UID: \"c577e6ac-c778-4859-8b39-30aeee31c0ec\") " pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.839678 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6ffr\" (UniqueName: \"kubernetes.io/projected/c577e6ac-c778-4859-8b39-30aeee31c0ec-kube-api-access-p6ffr\") pod \"metallb-operator-controller-manager-697d66d89d-jhzwv\" (UID: \"c577e6ac-c778-4859-8b39-30aeee31c0ec\") " pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.839738 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c577e6ac-c778-4859-8b39-30aeee31c0ec-apiservice-cert\") pod \"metallb-operator-controller-manager-697d66d89d-jhzwv\" (UID: \"c577e6ac-c778-4859-8b39-30aeee31c0ec\") " pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.909872 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj"] Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.910516 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.912640 4739 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.912671 4739 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.913041 4739 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-jwrxq" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.941243 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c577e6ac-c778-4859-8b39-30aeee31c0ec-apiservice-cert\") pod \"metallb-operator-controller-manager-697d66d89d-jhzwv\" (UID: \"c577e6ac-c778-4859-8b39-30aeee31c0ec\") " pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.941341 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c577e6ac-c778-4859-8b39-30aeee31c0ec-webhook-cert\") pod \"metallb-operator-controller-manager-697d66d89d-jhzwv\" (UID: \"c577e6ac-c778-4859-8b39-30aeee31c0ec\") " pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.941441 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6ffr\" (UniqueName: \"kubernetes.io/projected/c577e6ac-c778-4859-8b39-30aeee31c0ec-kube-api-access-p6ffr\") pod \"metallb-operator-controller-manager-697d66d89d-jhzwv\" (UID: \"c577e6ac-c778-4859-8b39-30aeee31c0ec\") " pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.954985 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c577e6ac-c778-4859-8b39-30aeee31c0ec-apiservice-cert\") pod \"metallb-operator-controller-manager-697d66d89d-jhzwv\" (UID: \"c577e6ac-c778-4859-8b39-30aeee31c0ec\") " pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.955396 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c577e6ac-c778-4859-8b39-30aeee31c0ec-webhook-cert\") pod \"metallb-operator-controller-manager-697d66d89d-jhzwv\" (UID: \"c577e6ac-c778-4859-8b39-30aeee31c0ec\") " pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.974601 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj"] Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.976573 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6ffr\" (UniqueName: \"kubernetes.io/projected/c577e6ac-c778-4859-8b39-30aeee31c0ec-kube-api-access-p6ffr\") pod \"metallb-operator-controller-manager-697d66d89d-jhzwv\" (UID: \"c577e6ac-c778-4859-8b39-30aeee31c0ec\") " pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" Dec 01 15:47:21 crc kubenswrapper[4739]: I1201 15:47:21.981764 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" Dec 01 15:47:22 crc kubenswrapper[4739]: I1201 15:47:22.044452 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d87d491-a2ce-4749-b3fd-8f429eeeda76-apiservice-cert\") pod \"metallb-operator-webhook-server-7cc9fc4f9f-67mlj\" (UID: \"1d87d491-a2ce-4749-b3fd-8f429eeeda76\") " pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" Dec 01 15:47:22 crc kubenswrapper[4739]: I1201 15:47:22.044848 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d87d491-a2ce-4749-b3fd-8f429eeeda76-webhook-cert\") pod \"metallb-operator-webhook-server-7cc9fc4f9f-67mlj\" (UID: \"1d87d491-a2ce-4749-b3fd-8f429eeeda76\") " pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" Dec 01 15:47:22 crc kubenswrapper[4739]: I1201 15:47:22.044974 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxsjv\" (UniqueName: \"kubernetes.io/projected/1d87d491-a2ce-4749-b3fd-8f429eeeda76-kube-api-access-vxsjv\") pod \"metallb-operator-webhook-server-7cc9fc4f9f-67mlj\" (UID: \"1d87d491-a2ce-4749-b3fd-8f429eeeda76\") " pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" Dec 01 15:47:22 crc kubenswrapper[4739]: I1201 15:47:22.148961 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxsjv\" (UniqueName: \"kubernetes.io/projected/1d87d491-a2ce-4749-b3fd-8f429eeeda76-kube-api-access-vxsjv\") pod \"metallb-operator-webhook-server-7cc9fc4f9f-67mlj\" (UID: \"1d87d491-a2ce-4749-b3fd-8f429eeeda76\") " pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" Dec 01 15:47:22 crc kubenswrapper[4739]: I1201 15:47:22.149026 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d87d491-a2ce-4749-b3fd-8f429eeeda76-apiservice-cert\") pod \"metallb-operator-webhook-server-7cc9fc4f9f-67mlj\" (UID: \"1d87d491-a2ce-4749-b3fd-8f429eeeda76\") " pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" Dec 01 15:47:22 crc kubenswrapper[4739]: I1201 15:47:22.149050 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d87d491-a2ce-4749-b3fd-8f429eeeda76-webhook-cert\") pod \"metallb-operator-webhook-server-7cc9fc4f9f-67mlj\" (UID: \"1d87d491-a2ce-4749-b3fd-8f429eeeda76\") " pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" Dec 01 15:47:22 crc kubenswrapper[4739]: I1201 15:47:22.159823 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d87d491-a2ce-4749-b3fd-8f429eeeda76-webhook-cert\") pod \"metallb-operator-webhook-server-7cc9fc4f9f-67mlj\" (UID: \"1d87d491-a2ce-4749-b3fd-8f429eeeda76\") " pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" Dec 01 15:47:22 crc kubenswrapper[4739]: I1201 15:47:22.165144 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d87d491-a2ce-4749-b3fd-8f429eeeda76-apiservice-cert\") pod \"metallb-operator-webhook-server-7cc9fc4f9f-67mlj\" (UID: \"1d87d491-a2ce-4749-b3fd-8f429eeeda76\") " pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" Dec 01 15:47:22 crc kubenswrapper[4739]: I1201 15:47:22.191470 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxsjv\" (UniqueName: \"kubernetes.io/projected/1d87d491-a2ce-4749-b3fd-8f429eeeda76-kube-api-access-vxsjv\") pod \"metallb-operator-webhook-server-7cc9fc4f9f-67mlj\" (UID: \"1d87d491-a2ce-4749-b3fd-8f429eeeda76\") " pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" Dec 01 15:47:22 crc kubenswrapper[4739]: I1201 15:47:22.223849 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" Dec 01 15:47:22 crc kubenswrapper[4739]: I1201 15:47:22.314900 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv"] Dec 01 15:47:22 crc kubenswrapper[4739]: W1201 15:47:22.325848 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc577e6ac_c778_4859_8b39_30aeee31c0ec.slice/crio-ea7e3efeeb9b1336c54aae5b5a4b6c3f75486749c63eae8941e64c7b289e5010 WatchSource:0}: Error finding container ea7e3efeeb9b1336c54aae5b5a4b6c3f75486749c63eae8941e64c7b289e5010: Status 404 returned error can't find the container with id ea7e3efeeb9b1336c54aae5b5a4b6c3f75486749c63eae8941e64c7b289e5010 Dec 01 15:47:22 crc kubenswrapper[4739]: I1201 15:47:22.500298 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj"] Dec 01 15:47:22 crc kubenswrapper[4739]: W1201 15:47:22.508408 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d87d491_a2ce_4749_b3fd_8f429eeeda76.slice/crio-e0d29596ce66178832439a9c7b870791541fc8a7514ae75fa235f689fdf22728 WatchSource:0}: Error finding container e0d29596ce66178832439a9c7b870791541fc8a7514ae75fa235f689fdf22728: Status 404 returned error can't find the container with id e0d29596ce66178832439a9c7b870791541fc8a7514ae75fa235f689fdf22728 Dec 01 15:47:22 crc kubenswrapper[4739]: I1201 15:47:22.742484 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" event={"ID":"1d87d491-a2ce-4749-b3fd-8f429eeeda76","Type":"ContainerStarted","Data":"e0d29596ce66178832439a9c7b870791541fc8a7514ae75fa235f689fdf22728"} Dec 01 15:47:22 crc kubenswrapper[4739]: I1201 15:47:22.743805 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" event={"ID":"c577e6ac-c778-4859-8b39-30aeee31c0ec","Type":"ContainerStarted","Data":"ea7e3efeeb9b1336c54aae5b5a4b6c3f75486749c63eae8941e64c7b289e5010"} Dec 01 15:47:24 crc kubenswrapper[4739]: I1201 15:47:24.762789 4739 generic.go:334] "Generic (PLEG): container finished" podID="c163b957-f79f-4bef-96d5-1a6d4353a849" containerID="fa32fb8c0dd993efe19ff16539b3817d2b405e6299e01cc8601822880e27a870" exitCode=0 Dec 01 15:47:24 crc kubenswrapper[4739]: I1201 15:47:24.762975 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfhc7" event={"ID":"c163b957-f79f-4bef-96d5-1a6d4353a849","Type":"ContainerDied","Data":"fa32fb8c0dd993efe19ff16539b3817d2b405e6299e01cc8601822880e27a870"} Dec 01 15:47:24 crc kubenswrapper[4739]: I1201 15:47:24.815360 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:24 crc kubenswrapper[4739]: I1201 15:47:24.913883 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c163b957-f79f-4bef-96d5-1a6d4353a849-utilities\") pod \"c163b957-f79f-4bef-96d5-1a6d4353a849\" (UID: \"c163b957-f79f-4bef-96d5-1a6d4353a849\") " Dec 01 15:47:24 crc kubenswrapper[4739]: I1201 15:47:24.913947 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c163b957-f79f-4bef-96d5-1a6d4353a849-catalog-content\") pod \"c163b957-f79f-4bef-96d5-1a6d4353a849\" (UID: \"c163b957-f79f-4bef-96d5-1a6d4353a849\") " Dec 01 15:47:24 crc kubenswrapper[4739]: I1201 15:47:24.913992 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kthdz\" (UniqueName: \"kubernetes.io/projected/c163b957-f79f-4bef-96d5-1a6d4353a849-kube-api-access-kthdz\") pod \"c163b957-f79f-4bef-96d5-1a6d4353a849\" (UID: \"c163b957-f79f-4bef-96d5-1a6d4353a849\") " Dec 01 15:47:24 crc kubenswrapper[4739]: I1201 15:47:24.915082 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c163b957-f79f-4bef-96d5-1a6d4353a849-utilities" (OuterVolumeSpecName: "utilities") pod "c163b957-f79f-4bef-96d5-1a6d4353a849" (UID: "c163b957-f79f-4bef-96d5-1a6d4353a849"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:47:24 crc kubenswrapper[4739]: I1201 15:47:24.920821 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c163b957-f79f-4bef-96d5-1a6d4353a849-kube-api-access-kthdz" (OuterVolumeSpecName: "kube-api-access-kthdz") pod "c163b957-f79f-4bef-96d5-1a6d4353a849" (UID: "c163b957-f79f-4bef-96d5-1a6d4353a849"). InnerVolumeSpecName "kube-api-access-kthdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:47:25 crc kubenswrapper[4739]: I1201 15:47:25.015898 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kthdz\" (UniqueName: \"kubernetes.io/projected/c163b957-f79f-4bef-96d5-1a6d4353a849-kube-api-access-kthdz\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:25 crc kubenswrapper[4739]: I1201 15:47:25.015961 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c163b957-f79f-4bef-96d5-1a6d4353a849-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:25 crc kubenswrapper[4739]: I1201 15:47:25.077112 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c163b957-f79f-4bef-96d5-1a6d4353a849-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c163b957-f79f-4bef-96d5-1a6d4353a849" (UID: "c163b957-f79f-4bef-96d5-1a6d4353a849"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:47:25 crc kubenswrapper[4739]: I1201 15:47:25.117778 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c163b957-f79f-4bef-96d5-1a6d4353a849-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:25 crc kubenswrapper[4739]: I1201 15:47:25.773741 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfhc7" event={"ID":"c163b957-f79f-4bef-96d5-1a6d4353a849","Type":"ContainerDied","Data":"93969a3070188f89d487d95ddb904ccee8575e4e8bf0997b88b38dd368ea0eb8"} Dec 01 15:47:25 crc kubenswrapper[4739]: I1201 15:47:25.774027 4739 scope.go:117] "RemoveContainer" containerID="fa32fb8c0dd993efe19ff16539b3817d2b405e6299e01cc8601822880e27a870" Dec 01 15:47:25 crc kubenswrapper[4739]: I1201 15:47:25.774135 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfhc7" Dec 01 15:47:25 crc kubenswrapper[4739]: I1201 15:47:25.808848 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bfhc7"] Dec 01 15:47:25 crc kubenswrapper[4739]: I1201 15:47:25.811680 4739 scope.go:117] "RemoveContainer" containerID="94d43bb96978f38ebcc6bffaa0e0f7f2f9a8cca4c76b0ca180feb5c3ea0404ee" Dec 01 15:47:25 crc kubenswrapper[4739]: I1201 15:47:25.812253 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bfhc7"] Dec 01 15:47:25 crc kubenswrapper[4739]: I1201 15:47:25.832021 4739 scope.go:117] "RemoveContainer" containerID="f203d52aac5e9a32dfe247f2f67905e162c7d2fadda454a8e0b23e8f756f2270" Dec 01 15:47:26 crc kubenswrapper[4739]: I1201 15:47:26.487652 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c163b957-f79f-4bef-96d5-1a6d4353a849" path="/var/lib/kubelet/pods/c163b957-f79f-4bef-96d5-1a6d4353a849/volumes" Dec 01 15:47:28 crc kubenswrapper[4739]: I1201 15:47:28.815010 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" event={"ID":"1d87d491-a2ce-4749-b3fd-8f429eeeda76","Type":"ContainerStarted","Data":"6772f1387c56658f1c430b54bc823b1f67d1ccdf39ab5ce4cad618a60d35a383"} Dec 01 15:47:28 crc kubenswrapper[4739]: I1201 15:47:28.815682 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" Dec 01 15:47:28 crc kubenswrapper[4739]: I1201 15:47:28.817049 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" event={"ID":"c577e6ac-c778-4859-8b39-30aeee31c0ec","Type":"ContainerStarted","Data":"1824a3a2c2db6d8de0ad3ab7eb8a0bc7a850661574ecd8645cfe15b9a4205f8c"} Dec 01 15:47:28 crc kubenswrapper[4739]: I1201 15:47:28.817228 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" Dec 01 15:47:28 crc kubenswrapper[4739]: I1201 15:47:28.836084 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" podStartSLOduration=1.988326724 podStartE2EDuration="7.836060587s" podCreationTimestamp="2025-12-01 15:47:21 +0000 UTC" firstStartedPulling="2025-12-01 15:47:22.511640393 +0000 UTC m=+744.337386487" lastFinishedPulling="2025-12-01 15:47:28.359374256 +0000 UTC m=+750.185120350" observedRunningTime="2025-12-01 15:47:28.830232869 +0000 UTC m=+750.655978963" watchObservedRunningTime="2025-12-01 15:47:28.836060587 +0000 UTC m=+750.661806691" Dec 01 15:47:28 crc kubenswrapper[4739]: I1201 15:47:28.862064 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" podStartSLOduration=1.8510814340000001 podStartE2EDuration="7.862045169s" podCreationTimestamp="2025-12-01 15:47:21 +0000 UTC" firstStartedPulling="2025-12-01 15:47:22.332810364 +0000 UTC m=+744.158556458" lastFinishedPulling="2025-12-01 15:47:28.343774059 +0000 UTC m=+750.169520193" observedRunningTime="2025-12-01 15:47:28.860173892 +0000 UTC m=+750.685920026" watchObservedRunningTime="2025-12-01 15:47:28.862045169 +0000 UTC m=+750.687791273" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.331089 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zzqk8"] Dec 01 15:47:33 crc kubenswrapper[4739]: E1201 15:47:33.332033 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c163b957-f79f-4bef-96d5-1a6d4353a849" containerName="registry-server" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.332094 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="c163b957-f79f-4bef-96d5-1a6d4353a849" containerName="registry-server" Dec 01 15:47:33 crc kubenswrapper[4739]: E1201 15:47:33.332141 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c163b957-f79f-4bef-96d5-1a6d4353a849" containerName="extract-content" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.332161 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="c163b957-f79f-4bef-96d5-1a6d4353a849" containerName="extract-content" Dec 01 15:47:33 crc kubenswrapper[4739]: E1201 15:47:33.332189 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c163b957-f79f-4bef-96d5-1a6d4353a849" containerName="extract-utilities" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.332204 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="c163b957-f79f-4bef-96d5-1a6d4353a849" containerName="extract-utilities" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.332513 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="c163b957-f79f-4bef-96d5-1a6d4353a849" containerName="registry-server" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.333946 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.346174 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zzqk8"] Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.438862 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlnvp\" (UniqueName: \"kubernetes.io/projected/1e809cb2-0750-470c-8a8b-a942fbb39128-kube-api-access-xlnvp\") pod \"redhat-marketplace-zzqk8\" (UID: \"1e809cb2-0750-470c-8a8b-a942fbb39128\") " pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.439103 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e809cb2-0750-470c-8a8b-a942fbb39128-catalog-content\") pod \"redhat-marketplace-zzqk8\" (UID: \"1e809cb2-0750-470c-8a8b-a942fbb39128\") " pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.439168 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e809cb2-0750-470c-8a8b-a942fbb39128-utilities\") pod \"redhat-marketplace-zzqk8\" (UID: \"1e809cb2-0750-470c-8a8b-a942fbb39128\") " pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.539957 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e809cb2-0750-470c-8a8b-a942fbb39128-catalog-content\") pod \"redhat-marketplace-zzqk8\" (UID: \"1e809cb2-0750-470c-8a8b-a942fbb39128\") " pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.540247 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e809cb2-0750-470c-8a8b-a942fbb39128-utilities\") pod \"redhat-marketplace-zzqk8\" (UID: \"1e809cb2-0750-470c-8a8b-a942fbb39128\") " pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.540404 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlnvp\" (UniqueName: \"kubernetes.io/projected/1e809cb2-0750-470c-8a8b-a942fbb39128-kube-api-access-xlnvp\") pod \"redhat-marketplace-zzqk8\" (UID: \"1e809cb2-0750-470c-8a8b-a942fbb39128\") " pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.540482 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e809cb2-0750-470c-8a8b-a942fbb39128-catalog-content\") pod \"redhat-marketplace-zzqk8\" (UID: \"1e809cb2-0750-470c-8a8b-a942fbb39128\") " pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.541023 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e809cb2-0750-470c-8a8b-a942fbb39128-utilities\") pod \"redhat-marketplace-zzqk8\" (UID: \"1e809cb2-0750-470c-8a8b-a942fbb39128\") " pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.569766 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlnvp\" (UniqueName: \"kubernetes.io/projected/1e809cb2-0750-470c-8a8b-a942fbb39128-kube-api-access-xlnvp\") pod \"redhat-marketplace-zzqk8\" (UID: \"1e809cb2-0750-470c-8a8b-a942fbb39128\") " pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.661313 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:33 crc kubenswrapper[4739]: I1201 15:47:33.862897 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zzqk8"] Dec 01 15:47:34 crc kubenswrapper[4739]: I1201 15:47:34.859353 4739 generic.go:334] "Generic (PLEG): container finished" podID="1e809cb2-0750-470c-8a8b-a942fbb39128" containerID="b6f5ddf61895fdc24e92514db3fe1c21a874820d42b203558b5da876055a0cf5" exitCode=0 Dec 01 15:47:34 crc kubenswrapper[4739]: I1201 15:47:34.859402 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zzqk8" event={"ID":"1e809cb2-0750-470c-8a8b-a942fbb39128","Type":"ContainerDied","Data":"b6f5ddf61895fdc24e92514db3fe1c21a874820d42b203558b5da876055a0cf5"} Dec 01 15:47:34 crc kubenswrapper[4739]: I1201 15:47:34.859747 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zzqk8" event={"ID":"1e809cb2-0750-470c-8a8b-a942fbb39128","Type":"ContainerStarted","Data":"821760e73918d90c2bb27b154e05a87a158273c13d2b384444d3829a671ff17f"} Dec 01 15:47:36 crc kubenswrapper[4739]: I1201 15:47:36.887379 4739 generic.go:334] "Generic (PLEG): container finished" podID="1e809cb2-0750-470c-8a8b-a942fbb39128" containerID="78f972fa5bdfec919719a34a502a7e1de71f829886c7450cd41669d5729aae70" exitCode=0 Dec 01 15:47:36 crc kubenswrapper[4739]: I1201 15:47:36.887446 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zzqk8" event={"ID":"1e809cb2-0750-470c-8a8b-a942fbb39128","Type":"ContainerDied","Data":"78f972fa5bdfec919719a34a502a7e1de71f829886c7450cd41669d5729aae70"} Dec 01 15:47:37 crc kubenswrapper[4739]: I1201 15:47:37.900933 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zzqk8" event={"ID":"1e809cb2-0750-470c-8a8b-a942fbb39128","Type":"ContainerStarted","Data":"4f9f45f3b7677086b92d60f05a046c1adca355a26677f9242457cbea33044b93"} Dec 01 15:47:39 crc kubenswrapper[4739]: I1201 15:47:39.621960 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:47:39 crc kubenswrapper[4739]: I1201 15:47:39.622311 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:47:39 crc kubenswrapper[4739]: I1201 15:47:39.622372 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:47:39 crc kubenswrapper[4739]: I1201 15:47:39.623186 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e3830d5cb4e555123caab94e2aaa75fd65f71b1d428fbec87bdc6d807609b113"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:47:39 crc kubenswrapper[4739]: I1201 15:47:39.623277 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://e3830d5cb4e555123caab94e2aaa75fd65f71b1d428fbec87bdc6d807609b113" gracePeriod=600 Dec 01 15:47:39 crc kubenswrapper[4739]: I1201 15:47:39.916831 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="e3830d5cb4e555123caab94e2aaa75fd65f71b1d428fbec87bdc6d807609b113" exitCode=0 Dec 01 15:47:39 crc kubenswrapper[4739]: I1201 15:47:39.916888 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"e3830d5cb4e555123caab94e2aaa75fd65f71b1d428fbec87bdc6d807609b113"} Dec 01 15:47:39 crc kubenswrapper[4739]: I1201 15:47:39.917100 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"c739ec8ed0d32f48e79459013f2a8032f453df5dc42b21acdeda98c2b463c62e"} Dec 01 15:47:39 crc kubenswrapper[4739]: I1201 15:47:39.917121 4739 scope.go:117] "RemoveContainer" containerID="d55c29fa9b125ad4bffaf3cb1c8640b7606b8326b206fa44f12ce15e676ba1b1" Dec 01 15:47:39 crc kubenswrapper[4739]: I1201 15:47:39.951410 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zzqk8" podStartSLOduration=4.201066128 podStartE2EDuration="6.951386152s" podCreationTimestamp="2025-12-01 15:47:33 +0000 UTC" firstStartedPulling="2025-12-01 15:47:34.860865373 +0000 UTC m=+756.686611467" lastFinishedPulling="2025-12-01 15:47:37.611185397 +0000 UTC m=+759.436931491" observedRunningTime="2025-12-01 15:47:37.925410349 +0000 UTC m=+759.751156443" watchObservedRunningTime="2025-12-01 15:47:39.951386152 +0000 UTC m=+761.777132276" Dec 01 15:47:42 crc kubenswrapper[4739]: I1201 15:47:42.229576 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7cc9fc4f9f-67mlj" Dec 01 15:47:43 crc kubenswrapper[4739]: I1201 15:47:43.662317 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:43 crc kubenswrapper[4739]: I1201 15:47:43.662735 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:43 crc kubenswrapper[4739]: I1201 15:47:43.718976 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:44 crc kubenswrapper[4739]: I1201 15:47:44.007962 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:46 crc kubenswrapper[4739]: I1201 15:47:46.107815 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zzqk8"] Dec 01 15:47:46 crc kubenswrapper[4739]: I1201 15:47:46.960497 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zzqk8" podUID="1e809cb2-0750-470c-8a8b-a942fbb39128" containerName="registry-server" containerID="cri-o://4f9f45f3b7677086b92d60f05a046c1adca355a26677f9242457cbea33044b93" gracePeriod=2 Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.371544 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.535800 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e809cb2-0750-470c-8a8b-a942fbb39128-utilities\") pod \"1e809cb2-0750-470c-8a8b-a942fbb39128\" (UID: \"1e809cb2-0750-470c-8a8b-a942fbb39128\") " Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.535846 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e809cb2-0750-470c-8a8b-a942fbb39128-catalog-content\") pod \"1e809cb2-0750-470c-8a8b-a942fbb39128\" (UID: \"1e809cb2-0750-470c-8a8b-a942fbb39128\") " Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.536161 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlnvp\" (UniqueName: \"kubernetes.io/projected/1e809cb2-0750-470c-8a8b-a942fbb39128-kube-api-access-xlnvp\") pod \"1e809cb2-0750-470c-8a8b-a942fbb39128\" (UID: \"1e809cb2-0750-470c-8a8b-a942fbb39128\") " Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.536764 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e809cb2-0750-470c-8a8b-a942fbb39128-utilities" (OuterVolumeSpecName: "utilities") pod "1e809cb2-0750-470c-8a8b-a942fbb39128" (UID: "1e809cb2-0750-470c-8a8b-a942fbb39128"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.536909 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e809cb2-0750-470c-8a8b-a942fbb39128-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.544407 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e809cb2-0750-470c-8a8b-a942fbb39128-kube-api-access-xlnvp" (OuterVolumeSpecName: "kube-api-access-xlnvp") pod "1e809cb2-0750-470c-8a8b-a942fbb39128" (UID: "1e809cb2-0750-470c-8a8b-a942fbb39128"). InnerVolumeSpecName "kube-api-access-xlnvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.567816 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e809cb2-0750-470c-8a8b-a942fbb39128-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e809cb2-0750-470c-8a8b-a942fbb39128" (UID: "1e809cb2-0750-470c-8a8b-a942fbb39128"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.637887 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlnvp\" (UniqueName: \"kubernetes.io/projected/1e809cb2-0750-470c-8a8b-a942fbb39128-kube-api-access-xlnvp\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.637937 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e809cb2-0750-470c-8a8b-a942fbb39128-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.967949 4739 generic.go:334] "Generic (PLEG): container finished" podID="1e809cb2-0750-470c-8a8b-a942fbb39128" containerID="4f9f45f3b7677086b92d60f05a046c1adca355a26677f9242457cbea33044b93" exitCode=0 Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.967998 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zzqk8" event={"ID":"1e809cb2-0750-470c-8a8b-a942fbb39128","Type":"ContainerDied","Data":"4f9f45f3b7677086b92d60f05a046c1adca355a26677f9242457cbea33044b93"} Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.968022 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zzqk8" Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.968047 4739 scope.go:117] "RemoveContainer" containerID="4f9f45f3b7677086b92d60f05a046c1adca355a26677f9242457cbea33044b93" Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.968031 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zzqk8" event={"ID":"1e809cb2-0750-470c-8a8b-a942fbb39128","Type":"ContainerDied","Data":"821760e73918d90c2bb27b154e05a87a158273c13d2b384444d3829a671ff17f"} Dec 01 15:47:47 crc kubenswrapper[4739]: I1201 15:47:47.990303 4739 scope.go:117] "RemoveContainer" containerID="78f972fa5bdfec919719a34a502a7e1de71f829886c7450cd41669d5729aae70" Dec 01 15:47:48 crc kubenswrapper[4739]: I1201 15:47:48.006325 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zzqk8"] Dec 01 15:47:48 crc kubenswrapper[4739]: I1201 15:47:48.024583 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zzqk8"] Dec 01 15:47:48 crc kubenswrapper[4739]: I1201 15:47:48.039911 4739 scope.go:117] "RemoveContainer" containerID="b6f5ddf61895fdc24e92514db3fe1c21a874820d42b203558b5da876055a0cf5" Dec 01 15:47:48 crc kubenswrapper[4739]: I1201 15:47:48.053519 4739 scope.go:117] "RemoveContainer" containerID="4f9f45f3b7677086b92d60f05a046c1adca355a26677f9242457cbea33044b93" Dec 01 15:47:48 crc kubenswrapper[4739]: E1201 15:47:48.053965 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f9f45f3b7677086b92d60f05a046c1adca355a26677f9242457cbea33044b93\": container with ID starting with 4f9f45f3b7677086b92d60f05a046c1adca355a26677f9242457cbea33044b93 not found: ID does not exist" containerID="4f9f45f3b7677086b92d60f05a046c1adca355a26677f9242457cbea33044b93" Dec 01 15:47:48 crc kubenswrapper[4739]: I1201 15:47:48.054010 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f9f45f3b7677086b92d60f05a046c1adca355a26677f9242457cbea33044b93"} err="failed to get container status \"4f9f45f3b7677086b92d60f05a046c1adca355a26677f9242457cbea33044b93\": rpc error: code = NotFound desc = could not find container \"4f9f45f3b7677086b92d60f05a046c1adca355a26677f9242457cbea33044b93\": container with ID starting with 4f9f45f3b7677086b92d60f05a046c1adca355a26677f9242457cbea33044b93 not found: ID does not exist" Dec 01 15:47:48 crc kubenswrapper[4739]: I1201 15:47:48.054036 4739 scope.go:117] "RemoveContainer" containerID="78f972fa5bdfec919719a34a502a7e1de71f829886c7450cd41669d5729aae70" Dec 01 15:47:48 crc kubenswrapper[4739]: E1201 15:47:48.054523 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78f972fa5bdfec919719a34a502a7e1de71f829886c7450cd41669d5729aae70\": container with ID starting with 78f972fa5bdfec919719a34a502a7e1de71f829886c7450cd41669d5729aae70 not found: ID does not exist" containerID="78f972fa5bdfec919719a34a502a7e1de71f829886c7450cd41669d5729aae70" Dec 01 15:47:48 crc kubenswrapper[4739]: I1201 15:47:48.054566 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78f972fa5bdfec919719a34a502a7e1de71f829886c7450cd41669d5729aae70"} err="failed to get container status \"78f972fa5bdfec919719a34a502a7e1de71f829886c7450cd41669d5729aae70\": rpc error: code = NotFound desc = could not find container \"78f972fa5bdfec919719a34a502a7e1de71f829886c7450cd41669d5729aae70\": container with ID starting with 78f972fa5bdfec919719a34a502a7e1de71f829886c7450cd41669d5729aae70 not found: ID does not exist" Dec 01 15:47:48 crc kubenswrapper[4739]: I1201 15:47:48.054595 4739 scope.go:117] "RemoveContainer" containerID="b6f5ddf61895fdc24e92514db3fe1c21a874820d42b203558b5da876055a0cf5" Dec 01 15:47:48 crc kubenswrapper[4739]: E1201 15:47:48.054966 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6f5ddf61895fdc24e92514db3fe1c21a874820d42b203558b5da876055a0cf5\": container with ID starting with b6f5ddf61895fdc24e92514db3fe1c21a874820d42b203558b5da876055a0cf5 not found: ID does not exist" containerID="b6f5ddf61895fdc24e92514db3fe1c21a874820d42b203558b5da876055a0cf5" Dec 01 15:47:48 crc kubenswrapper[4739]: I1201 15:47:48.055003 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6f5ddf61895fdc24e92514db3fe1c21a874820d42b203558b5da876055a0cf5"} err="failed to get container status \"b6f5ddf61895fdc24e92514db3fe1c21a874820d42b203558b5da876055a0cf5\": rpc error: code = NotFound desc = could not find container \"b6f5ddf61895fdc24e92514db3fe1c21a874820d42b203558b5da876055a0cf5\": container with ID starting with b6f5ddf61895fdc24e92514db3fe1c21a874820d42b203558b5da876055a0cf5 not found: ID does not exist" Dec 01 15:47:48 crc kubenswrapper[4739]: I1201 15:47:48.488789 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e809cb2-0750-470c-8a8b-a942fbb39128" path="/var/lib/kubelet/pods/1e809cb2-0750-470c-8a8b-a942fbb39128/volumes" Dec 01 15:48:01 crc kubenswrapper[4739]: I1201 15:48:01.985827 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-697d66d89d-jhzwv" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.750120 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-hzk9q"] Dec 01 15:48:02 crc kubenswrapper[4739]: E1201 15:48:02.750342 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e809cb2-0750-470c-8a8b-a942fbb39128" containerName="registry-server" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.750359 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e809cb2-0750-470c-8a8b-a942fbb39128" containerName="registry-server" Dec 01 15:48:02 crc kubenswrapper[4739]: E1201 15:48:02.750370 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e809cb2-0750-470c-8a8b-a942fbb39128" containerName="extract-content" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.750376 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e809cb2-0750-470c-8a8b-a942fbb39128" containerName="extract-content" Dec 01 15:48:02 crc kubenswrapper[4739]: E1201 15:48:02.750389 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e809cb2-0750-470c-8a8b-a942fbb39128" containerName="extract-utilities" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.750397 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e809cb2-0750-470c-8a8b-a942fbb39128" containerName="extract-utilities" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.750514 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e809cb2-0750-470c-8a8b-a942fbb39128" containerName="registry-server" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.752364 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.755393 4739 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.755467 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.755719 4739 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-g26k9" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.758044 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss"] Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.758943 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.764840 4739 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.787200 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss"] Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.935097 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b3db7f6-71d0-4951-89e8-da65693606e7-metrics-certs\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.935164 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pnbt\" (UniqueName: \"kubernetes.io/projected/d6492935-8d9c-4826-aa9d-26e4e27af9a0-kube-api-access-9pnbt\") pod \"frr-k8s-webhook-server-7fcb986d4-zs4ss\" (UID: \"d6492935-8d9c-4826-aa9d-26e4e27af9a0\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.935189 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6b3db7f6-71d0-4951-89e8-da65693606e7-metrics\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.935205 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d6492935-8d9c-4826-aa9d-26e4e27af9a0-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-zs4ss\" (UID: \"d6492935-8d9c-4826-aa9d-26e4e27af9a0\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.935228 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6b3db7f6-71d0-4951-89e8-da65693606e7-frr-startup\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.935242 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6b3db7f6-71d0-4951-89e8-da65693606e7-reloader\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.935259 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6b3db7f6-71d0-4951-89e8-da65693606e7-frr-sockets\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.935694 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cp8v\" (UniqueName: \"kubernetes.io/projected/6b3db7f6-71d0-4951-89e8-da65693606e7-kube-api-access-5cp8v\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.935851 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6b3db7f6-71d0-4951-89e8-da65693606e7-frr-conf\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.946227 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-cwjct"] Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.947260 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cwjct" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.950389 4739 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-2mhmn" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.950678 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.950787 4739 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.953828 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-t95pz"] Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.954733 4739 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.955016 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-t95pz" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.956931 4739 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 01 15:48:02 crc kubenswrapper[4739]: I1201 15:48:02.966859 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-t95pz"] Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.037464 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b3db7f6-71d0-4951-89e8-da65693606e7-metrics-certs\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.037516 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pnbt\" (UniqueName: \"kubernetes.io/projected/d6492935-8d9c-4826-aa9d-26e4e27af9a0-kube-api-access-9pnbt\") pod \"frr-k8s-webhook-server-7fcb986d4-zs4ss\" (UID: \"d6492935-8d9c-4826-aa9d-26e4e27af9a0\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.037545 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6b3db7f6-71d0-4951-89e8-da65693606e7-metrics\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.037566 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d6492935-8d9c-4826-aa9d-26e4e27af9a0-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-zs4ss\" (UID: \"d6492935-8d9c-4826-aa9d-26e4e27af9a0\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.037595 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6b3db7f6-71d0-4951-89e8-da65693606e7-frr-startup\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.037612 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6b3db7f6-71d0-4951-89e8-da65693606e7-reloader\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.037633 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f3469c68-6d37-41c5-bbca-32d8703b60e4-metrics-certs\") pod \"speaker-cwjct\" (UID: \"f3469c68-6d37-41c5-bbca-32d8703b60e4\") " pod="metallb-system/speaker-cwjct" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.037655 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6b3db7f6-71d0-4951-89e8-da65693606e7-frr-sockets\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.037680 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc5l2\" (UniqueName: \"kubernetes.io/projected/f3469c68-6d37-41c5-bbca-32d8703b60e4-kube-api-access-pc5l2\") pod \"speaker-cwjct\" (UID: \"f3469c68-6d37-41c5-bbca-32d8703b60e4\") " pod="metallb-system/speaker-cwjct" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.037696 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f3469c68-6d37-41c5-bbca-32d8703b60e4-metallb-excludel2\") pod \"speaker-cwjct\" (UID: \"f3469c68-6d37-41c5-bbca-32d8703b60e4\") " pod="metallb-system/speaker-cwjct" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.037715 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cp8v\" (UniqueName: \"kubernetes.io/projected/6b3db7f6-71d0-4951-89e8-da65693606e7-kube-api-access-5cp8v\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.037737 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b839c085-2a05-4b9a-940d-341fcaa4df18-cert\") pod \"controller-f8648f98b-t95pz\" (UID: \"b839c085-2a05-4b9a-940d-341fcaa4df18\") " pod="metallb-system/controller-f8648f98b-t95pz" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.037774 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6b3db7f6-71d0-4951-89e8-da65693606e7-frr-conf\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.037790 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f3469c68-6d37-41c5-bbca-32d8703b60e4-memberlist\") pod \"speaker-cwjct\" (UID: \"f3469c68-6d37-41c5-bbca-32d8703b60e4\") " pod="metallb-system/speaker-cwjct" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.038660 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6b3db7f6-71d0-4951-89e8-da65693606e7-metrics\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.038684 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6b3db7f6-71d0-4951-89e8-da65693606e7-reloader\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.038856 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6b3db7f6-71d0-4951-89e8-da65693606e7-frr-sockets\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.038894 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6b3db7f6-71d0-4951-89e8-da65693606e7-frr-conf\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.039555 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6b3db7f6-71d0-4951-89e8-da65693606e7-frr-startup\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.047153 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d6492935-8d9c-4826-aa9d-26e4e27af9a0-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-zs4ss\" (UID: \"d6492935-8d9c-4826-aa9d-26e4e27af9a0\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.047218 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b3db7f6-71d0-4951-89e8-da65693606e7-metrics-certs\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.056095 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pnbt\" (UniqueName: \"kubernetes.io/projected/d6492935-8d9c-4826-aa9d-26e4e27af9a0-kube-api-access-9pnbt\") pod \"frr-k8s-webhook-server-7fcb986d4-zs4ss\" (UID: \"d6492935-8d9c-4826-aa9d-26e4e27af9a0\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.066176 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cp8v\" (UniqueName: \"kubernetes.io/projected/6b3db7f6-71d0-4951-89e8-da65693606e7-kube-api-access-5cp8v\") pod \"frr-k8s-hzk9q\" (UID: \"6b3db7f6-71d0-4951-89e8-da65693606e7\") " pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.071921 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.139127 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b839c085-2a05-4b9a-940d-341fcaa4df18-metrics-certs\") pod \"controller-f8648f98b-t95pz\" (UID: \"b839c085-2a05-4b9a-940d-341fcaa4df18\") " pod="metallb-system/controller-f8648f98b-t95pz" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.139192 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f3469c68-6d37-41c5-bbca-32d8703b60e4-metrics-certs\") pod \"speaker-cwjct\" (UID: \"f3469c68-6d37-41c5-bbca-32d8703b60e4\") " pod="metallb-system/speaker-cwjct" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.139212 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlrbd\" (UniqueName: \"kubernetes.io/projected/b839c085-2a05-4b9a-940d-341fcaa4df18-kube-api-access-hlrbd\") pod \"controller-f8648f98b-t95pz\" (UID: \"b839c085-2a05-4b9a-940d-341fcaa4df18\") " pod="metallb-system/controller-f8648f98b-t95pz" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.139231 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc5l2\" (UniqueName: \"kubernetes.io/projected/f3469c68-6d37-41c5-bbca-32d8703b60e4-kube-api-access-pc5l2\") pod \"speaker-cwjct\" (UID: \"f3469c68-6d37-41c5-bbca-32d8703b60e4\") " pod="metallb-system/speaker-cwjct" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.139245 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f3469c68-6d37-41c5-bbca-32d8703b60e4-metallb-excludel2\") pod \"speaker-cwjct\" (UID: \"f3469c68-6d37-41c5-bbca-32d8703b60e4\") " pod="metallb-system/speaker-cwjct" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.139265 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b839c085-2a05-4b9a-940d-341fcaa4df18-cert\") pod \"controller-f8648f98b-t95pz\" (UID: \"b839c085-2a05-4b9a-940d-341fcaa4df18\") " pod="metallb-system/controller-f8648f98b-t95pz" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.139289 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f3469c68-6d37-41c5-bbca-32d8703b60e4-memberlist\") pod \"speaker-cwjct\" (UID: \"f3469c68-6d37-41c5-bbca-32d8703b60e4\") " pod="metallb-system/speaker-cwjct" Dec 01 15:48:03 crc kubenswrapper[4739]: E1201 15:48:03.139412 4739 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 15:48:03 crc kubenswrapper[4739]: E1201 15:48:03.139510 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f3469c68-6d37-41c5-bbca-32d8703b60e4-memberlist podName:f3469c68-6d37-41c5-bbca-32d8703b60e4 nodeName:}" failed. No retries permitted until 2025-12-01 15:48:03.639494932 +0000 UTC m=+785.465241026 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/f3469c68-6d37-41c5-bbca-32d8703b60e4-memberlist") pod "speaker-cwjct" (UID: "f3469c68-6d37-41c5-bbca-32d8703b60e4") : secret "metallb-memberlist" not found Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.140189 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f3469c68-6d37-41c5-bbca-32d8703b60e4-metallb-excludel2\") pod \"speaker-cwjct\" (UID: \"f3469c68-6d37-41c5-bbca-32d8703b60e4\") " pod="metallb-system/speaker-cwjct" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.141109 4739 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.143250 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f3469c68-6d37-41c5-bbca-32d8703b60e4-metrics-certs\") pod \"speaker-cwjct\" (UID: \"f3469c68-6d37-41c5-bbca-32d8703b60e4\") " pod="metallb-system/speaker-cwjct" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.157938 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b839c085-2a05-4b9a-940d-341fcaa4df18-cert\") pod \"controller-f8648f98b-t95pz\" (UID: \"b839c085-2a05-4b9a-940d-341fcaa4df18\") " pod="metallb-system/controller-f8648f98b-t95pz" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.160377 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc5l2\" (UniqueName: \"kubernetes.io/projected/f3469c68-6d37-41c5-bbca-32d8703b60e4-kube-api-access-pc5l2\") pod \"speaker-cwjct\" (UID: \"f3469c68-6d37-41c5-bbca-32d8703b60e4\") " pod="metallb-system/speaker-cwjct" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.160647 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.240411 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b839c085-2a05-4b9a-940d-341fcaa4df18-metrics-certs\") pod \"controller-f8648f98b-t95pz\" (UID: \"b839c085-2a05-4b9a-940d-341fcaa4df18\") " pod="metallb-system/controller-f8648f98b-t95pz" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.240483 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlrbd\" (UniqueName: \"kubernetes.io/projected/b839c085-2a05-4b9a-940d-341fcaa4df18-kube-api-access-hlrbd\") pod \"controller-f8648f98b-t95pz\" (UID: \"b839c085-2a05-4b9a-940d-341fcaa4df18\") " pod="metallb-system/controller-f8648f98b-t95pz" Dec 01 15:48:03 crc kubenswrapper[4739]: E1201 15:48:03.240888 4739 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 01 15:48:03 crc kubenswrapper[4739]: E1201 15:48:03.240986 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b839c085-2a05-4b9a-940d-341fcaa4df18-metrics-certs podName:b839c085-2a05-4b9a-940d-341fcaa4df18 nodeName:}" failed. No retries permitted until 2025-12-01 15:48:03.740944651 +0000 UTC m=+785.566690845 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b839c085-2a05-4b9a-940d-341fcaa4df18-metrics-certs") pod "controller-f8648f98b-t95pz" (UID: "b839c085-2a05-4b9a-940d-341fcaa4df18") : secret "controller-certs-secret" not found Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.261173 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlrbd\" (UniqueName: \"kubernetes.io/projected/b839c085-2a05-4b9a-940d-341fcaa4df18-kube-api-access-hlrbd\") pod \"controller-f8648f98b-t95pz\" (UID: \"b839c085-2a05-4b9a-940d-341fcaa4df18\") " pod="metallb-system/controller-f8648f98b-t95pz" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.365406 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss"] Dec 01 15:48:03 crc kubenswrapper[4739]: W1201 15:48:03.369228 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6492935_8d9c_4826_aa9d_26e4e27af9a0.slice/crio-4984bf5a70ca0f66fb1bb9af05e1934226537802db5b90625cc5c60cefb5c7b2 WatchSource:0}: Error finding container 4984bf5a70ca0f66fb1bb9af05e1934226537802db5b90625cc5c60cefb5c7b2: Status 404 returned error can't find the container with id 4984bf5a70ca0f66fb1bb9af05e1934226537802db5b90625cc5c60cefb5c7b2 Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.646935 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f3469c68-6d37-41c5-bbca-32d8703b60e4-memberlist\") pod \"speaker-cwjct\" (UID: \"f3469c68-6d37-41c5-bbca-32d8703b60e4\") " pod="metallb-system/speaker-cwjct" Dec 01 15:48:03 crc kubenswrapper[4739]: E1201 15:48:03.647171 4739 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 15:48:03 crc kubenswrapper[4739]: E1201 15:48:03.647298 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f3469c68-6d37-41c5-bbca-32d8703b60e4-memberlist podName:f3469c68-6d37-41c5-bbca-32d8703b60e4 nodeName:}" failed. No retries permitted until 2025-12-01 15:48:04.647259993 +0000 UTC m=+786.473006127 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/f3469c68-6d37-41c5-bbca-32d8703b60e4-memberlist") pod "speaker-cwjct" (UID: "f3469c68-6d37-41c5-bbca-32d8703b60e4") : secret "metallb-memberlist" not found Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.748510 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b839c085-2a05-4b9a-940d-341fcaa4df18-metrics-certs\") pod \"controller-f8648f98b-t95pz\" (UID: \"b839c085-2a05-4b9a-940d-341fcaa4df18\") " pod="metallb-system/controller-f8648f98b-t95pz" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.754053 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b839c085-2a05-4b9a-940d-341fcaa4df18-metrics-certs\") pod \"controller-f8648f98b-t95pz\" (UID: \"b839c085-2a05-4b9a-940d-341fcaa4df18\") " pod="metallb-system/controller-f8648f98b-t95pz" Dec 01 15:48:03 crc kubenswrapper[4739]: I1201 15:48:03.875484 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-t95pz" Dec 01 15:48:04 crc kubenswrapper[4739]: I1201 15:48:04.086193 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss" event={"ID":"d6492935-8d9c-4826-aa9d-26e4e27af9a0","Type":"ContainerStarted","Data":"4984bf5a70ca0f66fb1bb9af05e1934226537802db5b90625cc5c60cefb5c7b2"} Dec 01 15:48:04 crc kubenswrapper[4739]: I1201 15:48:04.087466 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hzk9q" event={"ID":"6b3db7f6-71d0-4951-89e8-da65693606e7","Type":"ContainerStarted","Data":"7233550f0984df21e97b7798d95b287ff4ac33046d0c8ec25c9cf05436ab5db7"} Dec 01 15:48:04 crc kubenswrapper[4739]: I1201 15:48:04.178971 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-t95pz"] Dec 01 15:48:04 crc kubenswrapper[4739]: I1201 15:48:04.660284 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f3469c68-6d37-41c5-bbca-32d8703b60e4-memberlist\") pod \"speaker-cwjct\" (UID: \"f3469c68-6d37-41c5-bbca-32d8703b60e4\") " pod="metallb-system/speaker-cwjct" Dec 01 15:48:04 crc kubenswrapper[4739]: E1201 15:48:04.660467 4739 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 15:48:04 crc kubenswrapper[4739]: E1201 15:48:04.660846 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f3469c68-6d37-41c5-bbca-32d8703b60e4-memberlist podName:f3469c68-6d37-41c5-bbca-32d8703b60e4 nodeName:}" failed. No retries permitted until 2025-12-01 15:48:06.660806155 +0000 UTC m=+788.486552269 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/f3469c68-6d37-41c5-bbca-32d8703b60e4-memberlist") pod "speaker-cwjct" (UID: "f3469c68-6d37-41c5-bbca-32d8703b60e4") : secret "metallb-memberlist" not found Dec 01 15:48:05 crc kubenswrapper[4739]: I1201 15:48:05.094252 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-t95pz" event={"ID":"b839c085-2a05-4b9a-940d-341fcaa4df18","Type":"ContainerStarted","Data":"5a1575dc75980c9368e448e4bee2a62162d6eeb602dc590a1fdca1fc7d756931"} Dec 01 15:48:05 crc kubenswrapper[4739]: I1201 15:48:05.094289 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-t95pz" event={"ID":"b839c085-2a05-4b9a-940d-341fcaa4df18","Type":"ContainerStarted","Data":"f5f6dff0c93497b279a60d462388f96d9c9925ed7c13296f536cb97e931d5301"} Dec 01 15:48:05 crc kubenswrapper[4739]: I1201 15:48:05.094298 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-t95pz" event={"ID":"b839c085-2a05-4b9a-940d-341fcaa4df18","Type":"ContainerStarted","Data":"faa70673c1eaaf60be6b0e1e16cd081d4584fbfdad4618d92f39aa4dd21c86a4"} Dec 01 15:48:05 crc kubenswrapper[4739]: I1201 15:48:05.094429 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-t95pz" Dec 01 15:48:05 crc kubenswrapper[4739]: I1201 15:48:05.114448 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-t95pz" podStartSLOduration=3.114407228 podStartE2EDuration="3.114407228s" podCreationTimestamp="2025-12-01 15:48:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:48:05.110030527 +0000 UTC m=+786.935776621" watchObservedRunningTime="2025-12-01 15:48:05.114407228 +0000 UTC m=+786.940153322" Dec 01 15:48:06 crc kubenswrapper[4739]: I1201 15:48:06.690352 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f3469c68-6d37-41c5-bbca-32d8703b60e4-memberlist\") pod \"speaker-cwjct\" (UID: \"f3469c68-6d37-41c5-bbca-32d8703b60e4\") " pod="metallb-system/speaker-cwjct" Dec 01 15:48:06 crc kubenswrapper[4739]: I1201 15:48:06.700231 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f3469c68-6d37-41c5-bbca-32d8703b60e4-memberlist\") pod \"speaker-cwjct\" (UID: \"f3469c68-6d37-41c5-bbca-32d8703b60e4\") " pod="metallb-system/speaker-cwjct" Dec 01 15:48:06 crc kubenswrapper[4739]: I1201 15:48:06.866840 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cwjct" Dec 01 15:48:06 crc kubenswrapper[4739]: W1201 15:48:06.886704 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3469c68_6d37_41c5_bbca_32d8703b60e4.slice/crio-18d31d7713c33411417f6cc708b38971c7e938a7dd720e43b79fd80ed28a5a51 WatchSource:0}: Error finding container 18d31d7713c33411417f6cc708b38971c7e938a7dd720e43b79fd80ed28a5a51: Status 404 returned error can't find the container with id 18d31d7713c33411417f6cc708b38971c7e938a7dd720e43b79fd80ed28a5a51 Dec 01 15:48:07 crc kubenswrapper[4739]: I1201 15:48:07.108050 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cwjct" event={"ID":"f3469c68-6d37-41c5-bbca-32d8703b60e4","Type":"ContainerStarted","Data":"18d31d7713c33411417f6cc708b38971c7e938a7dd720e43b79fd80ed28a5a51"} Dec 01 15:48:08 crc kubenswrapper[4739]: I1201 15:48:08.123557 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cwjct" event={"ID":"f3469c68-6d37-41c5-bbca-32d8703b60e4","Type":"ContainerStarted","Data":"7e757b1b9f0fa33d3016ac6a3648d726b2f7749ea357d551f0daf1bed5925e86"} Dec 01 15:48:08 crc kubenswrapper[4739]: I1201 15:48:08.123932 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-cwjct" Dec 01 15:48:08 crc kubenswrapper[4739]: I1201 15:48:08.123943 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cwjct" event={"ID":"f3469c68-6d37-41c5-bbca-32d8703b60e4","Type":"ContainerStarted","Data":"325a6a2736e5da9df8e8b3298fcd218a28f1d3d50e9e56401fdce67aabded36d"} Dec 01 15:48:08 crc kubenswrapper[4739]: I1201 15:48:08.508498 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-cwjct" podStartSLOduration=6.508477515 podStartE2EDuration="6.508477515s" podCreationTimestamp="2025-12-01 15:48:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:48:08.151564758 +0000 UTC m=+789.977310862" watchObservedRunningTime="2025-12-01 15:48:08.508477515 +0000 UTC m=+790.334223609" Dec 01 15:48:12 crc kubenswrapper[4739]: I1201 15:48:12.148486 4739 generic.go:334] "Generic (PLEG): container finished" podID="6b3db7f6-71d0-4951-89e8-da65693606e7" containerID="a5b233507ad6ad28255011acf42124e63cf44c07af907796e450e5c059e18e04" exitCode=0 Dec 01 15:48:12 crc kubenswrapper[4739]: I1201 15:48:12.148672 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hzk9q" event={"ID":"6b3db7f6-71d0-4951-89e8-da65693606e7","Type":"ContainerDied","Data":"a5b233507ad6ad28255011acf42124e63cf44c07af907796e450e5c059e18e04"} Dec 01 15:48:12 crc kubenswrapper[4739]: I1201 15:48:12.151491 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss" event={"ID":"d6492935-8d9c-4826-aa9d-26e4e27af9a0","Type":"ContainerStarted","Data":"67e88b4077894253aea03a6bee8c901c3ee1594095e9d1872d66008f916fb230"} Dec 01 15:48:12 crc kubenswrapper[4739]: I1201 15:48:12.151657 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss" Dec 01 15:48:12 crc kubenswrapper[4739]: I1201 15:48:12.208876 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss" podStartSLOduration=2.518706684 podStartE2EDuration="10.208842398s" podCreationTimestamp="2025-12-01 15:48:02 +0000 UTC" firstStartedPulling="2025-12-01 15:48:03.370502855 +0000 UTC m=+785.196248949" lastFinishedPulling="2025-12-01 15:48:11.060638559 +0000 UTC m=+792.886384663" observedRunningTime="2025-12-01 15:48:12.202324771 +0000 UTC m=+794.028070885" watchObservedRunningTime="2025-12-01 15:48:12.208842398 +0000 UTC m=+794.034588532" Dec 01 15:48:13 crc kubenswrapper[4739]: I1201 15:48:13.164139 4739 generic.go:334] "Generic (PLEG): container finished" podID="6b3db7f6-71d0-4951-89e8-da65693606e7" containerID="57de36f030eb25320e7b10249fc3940036de6b54abf59f9c89daf6a0d69333dc" exitCode=0 Dec 01 15:48:13 crc kubenswrapper[4739]: I1201 15:48:13.164206 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hzk9q" event={"ID":"6b3db7f6-71d0-4951-89e8-da65693606e7","Type":"ContainerDied","Data":"57de36f030eb25320e7b10249fc3940036de6b54abf59f9c89daf6a0d69333dc"} Dec 01 15:48:14 crc kubenswrapper[4739]: I1201 15:48:14.177014 4739 generic.go:334] "Generic (PLEG): container finished" podID="6b3db7f6-71d0-4951-89e8-da65693606e7" containerID="24137ba81fa0697980a5c273c4dd7c3f493af14411533c2dd1de5cfcf6fb251d" exitCode=0 Dec 01 15:48:14 crc kubenswrapper[4739]: I1201 15:48:14.177082 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hzk9q" event={"ID":"6b3db7f6-71d0-4951-89e8-da65693606e7","Type":"ContainerDied","Data":"24137ba81fa0697980a5c273c4dd7c3f493af14411533c2dd1de5cfcf6fb251d"} Dec 01 15:48:15 crc kubenswrapper[4739]: I1201 15:48:15.187765 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hzk9q" event={"ID":"6b3db7f6-71d0-4951-89e8-da65693606e7","Type":"ContainerStarted","Data":"d97a3dce6f559295e217387ea9dcb0ebb1ad14987c40cf9a7f95177b577c1e8d"} Dec 01 15:48:15 crc kubenswrapper[4739]: I1201 15:48:15.188229 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hzk9q" event={"ID":"6b3db7f6-71d0-4951-89e8-da65693606e7","Type":"ContainerStarted","Data":"d1a93b67664a11e9ea7392765155d8a4b33b3f4a4225f9b307867e26576035f4"} Dec 01 15:48:15 crc kubenswrapper[4739]: I1201 15:48:15.188245 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hzk9q" event={"ID":"6b3db7f6-71d0-4951-89e8-da65693606e7","Type":"ContainerStarted","Data":"6eab3463f84828ed8cacf16425ac5d8734dd6fa4e0eada55ef6a32bcbf234ce7"} Dec 01 15:48:15 crc kubenswrapper[4739]: I1201 15:48:15.188255 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hzk9q" event={"ID":"6b3db7f6-71d0-4951-89e8-da65693606e7","Type":"ContainerStarted","Data":"28498c1640ac5ff42308b78ede30f2e82425a4089b8eea0b33b3df62344b5e7d"} Dec 01 15:48:15 crc kubenswrapper[4739]: I1201 15:48:15.188263 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hzk9q" event={"ID":"6b3db7f6-71d0-4951-89e8-da65693606e7","Type":"ContainerStarted","Data":"97f13c3df4ca11761fc75b7dfb7af5ad0f3bd941d06f6e6e39d988ab1c09eaf8"} Dec 01 15:48:16 crc kubenswrapper[4739]: I1201 15:48:16.199998 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hzk9q" event={"ID":"6b3db7f6-71d0-4951-89e8-da65693606e7","Type":"ContainerStarted","Data":"558ba7a7a0c05ae492b5ea047784b9f91111f21af6f47fb5f60e268d55ee5ad6"} Dec 01 15:48:16 crc kubenswrapper[4739]: I1201 15:48:16.200814 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:16 crc kubenswrapper[4739]: I1201 15:48:16.238323 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-hzk9q" podStartSLOduration=6.815678137 podStartE2EDuration="14.238298621s" podCreationTimestamp="2025-12-01 15:48:02 +0000 UTC" firstStartedPulling="2025-12-01 15:48:03.613080815 +0000 UTC m=+785.438826949" lastFinishedPulling="2025-12-01 15:48:11.035701339 +0000 UTC m=+792.861447433" observedRunningTime="2025-12-01 15:48:16.232546948 +0000 UTC m=+798.058293082" watchObservedRunningTime="2025-12-01 15:48:16.238298621 +0000 UTC m=+798.064044745" Dec 01 15:48:18 crc kubenswrapper[4739]: I1201 15:48:18.073326 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:18 crc kubenswrapper[4739]: I1201 15:48:18.109023 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:23 crc kubenswrapper[4739]: I1201 15:48:23.170921 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zs4ss" Dec 01 15:48:23 crc kubenswrapper[4739]: I1201 15:48:23.882587 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-t95pz" Dec 01 15:48:26 crc kubenswrapper[4739]: I1201 15:48:26.872123 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-cwjct" Dec 01 15:48:29 crc kubenswrapper[4739]: I1201 15:48:29.604712 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zwb67"] Dec 01 15:48:29 crc kubenswrapper[4739]: I1201 15:48:29.605891 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zwb67" Dec 01 15:48:29 crc kubenswrapper[4739]: I1201 15:48:29.608636 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 01 15:48:29 crc kubenswrapper[4739]: I1201 15:48:29.609818 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 01 15:48:29 crc kubenswrapper[4739]: I1201 15:48:29.614200 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-fnhpv" Dec 01 15:48:29 crc kubenswrapper[4739]: I1201 15:48:29.624282 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7xfv\" (UniqueName: \"kubernetes.io/projected/70f7e5af-1723-4f90-8514-8e0495da984d-kube-api-access-h7xfv\") pod \"openstack-operator-index-zwb67\" (UID: \"70f7e5af-1723-4f90-8514-8e0495da984d\") " pod="openstack-operators/openstack-operator-index-zwb67" Dec 01 15:48:29 crc kubenswrapper[4739]: I1201 15:48:29.627446 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zwb67"] Dec 01 15:48:29 crc kubenswrapper[4739]: I1201 15:48:29.725675 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7xfv\" (UniqueName: \"kubernetes.io/projected/70f7e5af-1723-4f90-8514-8e0495da984d-kube-api-access-h7xfv\") pod \"openstack-operator-index-zwb67\" (UID: \"70f7e5af-1723-4f90-8514-8e0495da984d\") " pod="openstack-operators/openstack-operator-index-zwb67" Dec 01 15:48:29 crc kubenswrapper[4739]: I1201 15:48:29.745356 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7xfv\" (UniqueName: \"kubernetes.io/projected/70f7e5af-1723-4f90-8514-8e0495da984d-kube-api-access-h7xfv\") pod \"openstack-operator-index-zwb67\" (UID: \"70f7e5af-1723-4f90-8514-8e0495da984d\") " pod="openstack-operators/openstack-operator-index-zwb67" Dec 01 15:48:29 crc kubenswrapper[4739]: I1201 15:48:29.921290 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zwb67" Dec 01 15:48:30 crc kubenswrapper[4739]: I1201 15:48:30.346925 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zwb67"] Dec 01 15:48:30 crc kubenswrapper[4739]: W1201 15:48:30.352351 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70f7e5af_1723_4f90_8514_8e0495da984d.slice/crio-9ad3088633397a0aba825d595c35bf2e799c336efe3572c0168792a0ec3d2b1c WatchSource:0}: Error finding container 9ad3088633397a0aba825d595c35bf2e799c336efe3572c0168792a0ec3d2b1c: Status 404 returned error can't find the container with id 9ad3088633397a0aba825d595c35bf2e799c336efe3572c0168792a0ec3d2b1c Dec 01 15:48:31 crc kubenswrapper[4739]: I1201 15:48:31.327074 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zwb67" event={"ID":"70f7e5af-1723-4f90-8514-8e0495da984d","Type":"ContainerStarted","Data":"9ad3088633397a0aba825d595c35bf2e799c336efe3572c0168792a0ec3d2b1c"} Dec 01 15:48:33 crc kubenswrapper[4739]: I1201 15:48:33.076196 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-hzk9q" Dec 01 15:48:33 crc kubenswrapper[4739]: I1201 15:48:33.541280 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-zwb67"] Dec 01 15:48:34 crc kubenswrapper[4739]: I1201 15:48:34.158818 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-sqnf9"] Dec 01 15:48:34 crc kubenswrapper[4739]: I1201 15:48:34.160774 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-sqnf9" Dec 01 15:48:34 crc kubenswrapper[4739]: I1201 15:48:34.169112 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-sqnf9"] Dec 01 15:48:34 crc kubenswrapper[4739]: I1201 15:48:34.286669 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxlzx\" (UniqueName: \"kubernetes.io/projected/38434f68-049c-4b26-ad64-c6da629a6df2-kube-api-access-nxlzx\") pod \"openstack-operator-index-sqnf9\" (UID: \"38434f68-049c-4b26-ad64-c6da629a6df2\") " pod="openstack-operators/openstack-operator-index-sqnf9" Dec 01 15:48:34 crc kubenswrapper[4739]: I1201 15:48:34.388865 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxlzx\" (UniqueName: \"kubernetes.io/projected/38434f68-049c-4b26-ad64-c6da629a6df2-kube-api-access-nxlzx\") pod \"openstack-operator-index-sqnf9\" (UID: \"38434f68-049c-4b26-ad64-c6da629a6df2\") " pod="openstack-operators/openstack-operator-index-sqnf9" Dec 01 15:48:34 crc kubenswrapper[4739]: I1201 15:48:34.419380 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxlzx\" (UniqueName: \"kubernetes.io/projected/38434f68-049c-4b26-ad64-c6da629a6df2-kube-api-access-nxlzx\") pod \"openstack-operator-index-sqnf9\" (UID: \"38434f68-049c-4b26-ad64-c6da629a6df2\") " pod="openstack-operators/openstack-operator-index-sqnf9" Dec 01 15:48:34 crc kubenswrapper[4739]: I1201 15:48:34.537863 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-sqnf9" Dec 01 15:48:35 crc kubenswrapper[4739]: I1201 15:48:35.695245 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-sqnf9"] Dec 01 15:48:36 crc kubenswrapper[4739]: I1201 15:48:36.374903 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-sqnf9" event={"ID":"38434f68-049c-4b26-ad64-c6da629a6df2","Type":"ContainerStarted","Data":"b9d425489f8de3c4ff05e27610e2fd9598b2b2f203ab8fdc940b945036b2ea1f"} Dec 01 15:48:38 crc kubenswrapper[4739]: I1201 15:48:38.388383 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-sqnf9" event={"ID":"38434f68-049c-4b26-ad64-c6da629a6df2","Type":"ContainerStarted","Data":"04218145f1866882a50403bb696be41f577209c7dc424b704ae9afb9e776850f"} Dec 01 15:48:38 crc kubenswrapper[4739]: I1201 15:48:38.389935 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zwb67" event={"ID":"70f7e5af-1723-4f90-8514-8e0495da984d","Type":"ContainerStarted","Data":"461d6ec55993a52ac4b8acf53a941f25aa9bba04b4eb8916cbf38177dcbb0f16"} Dec 01 15:48:38 crc kubenswrapper[4739]: I1201 15:48:38.390081 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-zwb67" podUID="70f7e5af-1723-4f90-8514-8e0495da984d" containerName="registry-server" containerID="cri-o://461d6ec55993a52ac4b8acf53a941f25aa9bba04b4eb8916cbf38177dcbb0f16" gracePeriod=2 Dec 01 15:48:38 crc kubenswrapper[4739]: I1201 15:48:38.407148 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-sqnf9" podStartSLOduration=2.600869307 podStartE2EDuration="4.407127024s" podCreationTimestamp="2025-12-01 15:48:34 +0000 UTC" firstStartedPulling="2025-12-01 15:48:35.711122056 +0000 UTC m=+817.536868150" lastFinishedPulling="2025-12-01 15:48:37.517379783 +0000 UTC m=+819.343125867" observedRunningTime="2025-12-01 15:48:38.406174785 +0000 UTC m=+820.231920919" watchObservedRunningTime="2025-12-01 15:48:38.407127024 +0000 UTC m=+820.232873118" Dec 01 15:48:38 crc kubenswrapper[4739]: I1201 15:48:38.430728 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zwb67" podStartSLOduration=2.271136425 podStartE2EDuration="9.430703642s" podCreationTimestamp="2025-12-01 15:48:29 +0000 UTC" firstStartedPulling="2025-12-01 15:48:30.355791055 +0000 UTC m=+812.181537159" lastFinishedPulling="2025-12-01 15:48:37.515358242 +0000 UTC m=+819.341104376" observedRunningTime="2025-12-01 15:48:38.425855197 +0000 UTC m=+820.251601291" watchObservedRunningTime="2025-12-01 15:48:38.430703642 +0000 UTC m=+820.256449766" Dec 01 15:48:38 crc kubenswrapper[4739]: I1201 15:48:38.835444 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zwb67" Dec 01 15:48:38 crc kubenswrapper[4739]: I1201 15:48:38.952343 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7xfv\" (UniqueName: \"kubernetes.io/projected/70f7e5af-1723-4f90-8514-8e0495da984d-kube-api-access-h7xfv\") pod \"70f7e5af-1723-4f90-8514-8e0495da984d\" (UID: \"70f7e5af-1723-4f90-8514-8e0495da984d\") " Dec 01 15:48:38 crc kubenswrapper[4739]: I1201 15:48:38.960207 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70f7e5af-1723-4f90-8514-8e0495da984d-kube-api-access-h7xfv" (OuterVolumeSpecName: "kube-api-access-h7xfv") pod "70f7e5af-1723-4f90-8514-8e0495da984d" (UID: "70f7e5af-1723-4f90-8514-8e0495da984d"). InnerVolumeSpecName "kube-api-access-h7xfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:48:39 crc kubenswrapper[4739]: I1201 15:48:39.053959 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7xfv\" (UniqueName: \"kubernetes.io/projected/70f7e5af-1723-4f90-8514-8e0495da984d-kube-api-access-h7xfv\") on node \"crc\" DevicePath \"\"" Dec 01 15:48:39 crc kubenswrapper[4739]: I1201 15:48:39.400917 4739 generic.go:334] "Generic (PLEG): container finished" podID="70f7e5af-1723-4f90-8514-8e0495da984d" containerID="461d6ec55993a52ac4b8acf53a941f25aa9bba04b4eb8916cbf38177dcbb0f16" exitCode=0 Dec 01 15:48:39 crc kubenswrapper[4739]: I1201 15:48:39.401008 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zwb67" event={"ID":"70f7e5af-1723-4f90-8514-8e0495da984d","Type":"ContainerDied","Data":"461d6ec55993a52ac4b8acf53a941f25aa9bba04b4eb8916cbf38177dcbb0f16"} Dec 01 15:48:39 crc kubenswrapper[4739]: I1201 15:48:39.401062 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zwb67" Dec 01 15:48:39 crc kubenswrapper[4739]: I1201 15:48:39.401090 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zwb67" event={"ID":"70f7e5af-1723-4f90-8514-8e0495da984d","Type":"ContainerDied","Data":"9ad3088633397a0aba825d595c35bf2e799c336efe3572c0168792a0ec3d2b1c"} Dec 01 15:48:39 crc kubenswrapper[4739]: I1201 15:48:39.401122 4739 scope.go:117] "RemoveContainer" containerID="461d6ec55993a52ac4b8acf53a941f25aa9bba04b4eb8916cbf38177dcbb0f16" Dec 01 15:48:39 crc kubenswrapper[4739]: I1201 15:48:39.440262 4739 scope.go:117] "RemoveContainer" containerID="461d6ec55993a52ac4b8acf53a941f25aa9bba04b4eb8916cbf38177dcbb0f16" Dec 01 15:48:39 crc kubenswrapper[4739]: E1201 15:48:39.440950 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"461d6ec55993a52ac4b8acf53a941f25aa9bba04b4eb8916cbf38177dcbb0f16\": container with ID starting with 461d6ec55993a52ac4b8acf53a941f25aa9bba04b4eb8916cbf38177dcbb0f16 not found: ID does not exist" containerID="461d6ec55993a52ac4b8acf53a941f25aa9bba04b4eb8916cbf38177dcbb0f16" Dec 01 15:48:39 crc kubenswrapper[4739]: I1201 15:48:39.441003 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"461d6ec55993a52ac4b8acf53a941f25aa9bba04b4eb8916cbf38177dcbb0f16"} err="failed to get container status \"461d6ec55993a52ac4b8acf53a941f25aa9bba04b4eb8916cbf38177dcbb0f16\": rpc error: code = NotFound desc = could not find container \"461d6ec55993a52ac4b8acf53a941f25aa9bba04b4eb8916cbf38177dcbb0f16\": container with ID starting with 461d6ec55993a52ac4b8acf53a941f25aa9bba04b4eb8916cbf38177dcbb0f16 not found: ID does not exist" Dec 01 15:48:39 crc kubenswrapper[4739]: I1201 15:48:39.447634 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-zwb67"] Dec 01 15:48:39 crc kubenswrapper[4739]: I1201 15:48:39.454969 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-zwb67"] Dec 01 15:48:40 crc kubenswrapper[4739]: I1201 15:48:40.484553 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70f7e5af-1723-4f90-8514-8e0495da984d" path="/var/lib/kubelet/pods/70f7e5af-1723-4f90-8514-8e0495da984d/volumes" Dec 01 15:48:44 crc kubenswrapper[4739]: I1201 15:48:44.538167 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-sqnf9" Dec 01 15:48:44 crc kubenswrapper[4739]: I1201 15:48:44.538573 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-sqnf9" Dec 01 15:48:44 crc kubenswrapper[4739]: I1201 15:48:44.580937 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-sqnf9" Dec 01 15:48:45 crc kubenswrapper[4739]: I1201 15:48:45.487453 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-sqnf9" Dec 01 15:48:53 crc kubenswrapper[4739]: I1201 15:48:53.463888 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp"] Dec 01 15:48:53 crc kubenswrapper[4739]: E1201 15:48:53.465215 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f7e5af-1723-4f90-8514-8e0495da984d" containerName="registry-server" Dec 01 15:48:53 crc kubenswrapper[4739]: I1201 15:48:53.465239 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f7e5af-1723-4f90-8514-8e0495da984d" containerName="registry-server" Dec 01 15:48:53 crc kubenswrapper[4739]: I1201 15:48:53.465465 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f7e5af-1723-4f90-8514-8e0495da984d" containerName="registry-server" Dec 01 15:48:53 crc kubenswrapper[4739]: I1201 15:48:53.466889 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" Dec 01 15:48:53 crc kubenswrapper[4739]: I1201 15:48:53.468781 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-9dbc7" Dec 01 15:48:53 crc kubenswrapper[4739]: I1201 15:48:53.472902 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp"] Dec 01 15:48:53 crc kubenswrapper[4739]: I1201 15:48:53.668791 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/500c937c-6baf-47e7-93ec-fbff3d5b6d37-bundle\") pod \"b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp\" (UID: \"500c937c-6baf-47e7-93ec-fbff3d5b6d37\") " pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" Dec 01 15:48:53 crc kubenswrapper[4739]: I1201 15:48:53.668857 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/500c937c-6baf-47e7-93ec-fbff3d5b6d37-util\") pod \"b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp\" (UID: \"500c937c-6baf-47e7-93ec-fbff3d5b6d37\") " pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" Dec 01 15:48:53 crc kubenswrapper[4739]: I1201 15:48:53.668886 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22g24\" (UniqueName: \"kubernetes.io/projected/500c937c-6baf-47e7-93ec-fbff3d5b6d37-kube-api-access-22g24\") pod \"b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp\" (UID: \"500c937c-6baf-47e7-93ec-fbff3d5b6d37\") " pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" Dec 01 15:48:53 crc kubenswrapper[4739]: I1201 15:48:53.769650 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/500c937c-6baf-47e7-93ec-fbff3d5b6d37-bundle\") pod \"b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp\" (UID: \"500c937c-6baf-47e7-93ec-fbff3d5b6d37\") " pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" Dec 01 15:48:53 crc kubenswrapper[4739]: I1201 15:48:53.769698 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/500c937c-6baf-47e7-93ec-fbff3d5b6d37-util\") pod \"b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp\" (UID: \"500c937c-6baf-47e7-93ec-fbff3d5b6d37\") " pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" Dec 01 15:48:53 crc kubenswrapper[4739]: I1201 15:48:53.769720 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22g24\" (UniqueName: \"kubernetes.io/projected/500c937c-6baf-47e7-93ec-fbff3d5b6d37-kube-api-access-22g24\") pod \"b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp\" (UID: \"500c937c-6baf-47e7-93ec-fbff3d5b6d37\") " pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" Dec 01 15:48:53 crc kubenswrapper[4739]: I1201 15:48:53.770219 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/500c937c-6baf-47e7-93ec-fbff3d5b6d37-bundle\") pod \"b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp\" (UID: \"500c937c-6baf-47e7-93ec-fbff3d5b6d37\") " pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" Dec 01 15:48:53 crc kubenswrapper[4739]: I1201 15:48:53.771165 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/500c937c-6baf-47e7-93ec-fbff3d5b6d37-util\") pod \"b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp\" (UID: \"500c937c-6baf-47e7-93ec-fbff3d5b6d37\") " pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" Dec 01 15:48:53 crc kubenswrapper[4739]: I1201 15:48:53.803472 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22g24\" (UniqueName: \"kubernetes.io/projected/500c937c-6baf-47e7-93ec-fbff3d5b6d37-kube-api-access-22g24\") pod \"b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp\" (UID: \"500c937c-6baf-47e7-93ec-fbff3d5b6d37\") " pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" Dec 01 15:48:54 crc kubenswrapper[4739]: I1201 15:48:54.093455 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" Dec 01 15:48:54 crc kubenswrapper[4739]: I1201 15:48:54.522998 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp"] Dec 01 15:48:55 crc kubenswrapper[4739]: I1201 15:48:55.535411 4739 generic.go:334] "Generic (PLEG): container finished" podID="500c937c-6baf-47e7-93ec-fbff3d5b6d37" containerID="cbbcb8361e3f26abc61bfef12a42eb73f593bae88abbcde5b25198008e7fa669" exitCode=0 Dec 01 15:48:55 crc kubenswrapper[4739]: I1201 15:48:55.535698 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" event={"ID":"500c937c-6baf-47e7-93ec-fbff3d5b6d37","Type":"ContainerDied","Data":"cbbcb8361e3f26abc61bfef12a42eb73f593bae88abbcde5b25198008e7fa669"} Dec 01 15:48:55 crc kubenswrapper[4739]: I1201 15:48:55.535853 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" event={"ID":"500c937c-6baf-47e7-93ec-fbff3d5b6d37","Type":"ContainerStarted","Data":"168485495e0c2b882b64308b350d9db53d815a0ac00b708fcb30c92e1f8d365d"} Dec 01 15:48:56 crc kubenswrapper[4739]: I1201 15:48:56.545747 4739 generic.go:334] "Generic (PLEG): container finished" podID="500c937c-6baf-47e7-93ec-fbff3d5b6d37" containerID="ab7e318fab8a21227b95b1f3907e23ca2980c495eb3c19d1e147e90c9f029292" exitCode=0 Dec 01 15:48:56 crc kubenswrapper[4739]: I1201 15:48:56.545811 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" event={"ID":"500c937c-6baf-47e7-93ec-fbff3d5b6d37","Type":"ContainerDied","Data":"ab7e318fab8a21227b95b1f3907e23ca2980c495eb3c19d1e147e90c9f029292"} Dec 01 15:48:57 crc kubenswrapper[4739]: I1201 15:48:57.556115 4739 generic.go:334] "Generic (PLEG): container finished" podID="500c937c-6baf-47e7-93ec-fbff3d5b6d37" containerID="e54b035645c479c9bd05c1510c1000ff7d35695719f469662f5da6fcf965c6c4" exitCode=0 Dec 01 15:48:57 crc kubenswrapper[4739]: I1201 15:48:57.556154 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" event={"ID":"500c937c-6baf-47e7-93ec-fbff3d5b6d37","Type":"ContainerDied","Data":"e54b035645c479c9bd05c1510c1000ff7d35695719f469662f5da6fcf965c6c4"} Dec 01 15:48:58 crc kubenswrapper[4739]: I1201 15:48:58.933466 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" Dec 01 15:48:59 crc kubenswrapper[4739]: I1201 15:48:59.040257 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/500c937c-6baf-47e7-93ec-fbff3d5b6d37-bundle\") pod \"500c937c-6baf-47e7-93ec-fbff3d5b6d37\" (UID: \"500c937c-6baf-47e7-93ec-fbff3d5b6d37\") " Dec 01 15:48:59 crc kubenswrapper[4739]: I1201 15:48:59.040361 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/500c937c-6baf-47e7-93ec-fbff3d5b6d37-util\") pod \"500c937c-6baf-47e7-93ec-fbff3d5b6d37\" (UID: \"500c937c-6baf-47e7-93ec-fbff3d5b6d37\") " Dec 01 15:48:59 crc kubenswrapper[4739]: I1201 15:48:59.040501 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22g24\" (UniqueName: \"kubernetes.io/projected/500c937c-6baf-47e7-93ec-fbff3d5b6d37-kube-api-access-22g24\") pod \"500c937c-6baf-47e7-93ec-fbff3d5b6d37\" (UID: \"500c937c-6baf-47e7-93ec-fbff3d5b6d37\") " Dec 01 15:48:59 crc kubenswrapper[4739]: I1201 15:48:59.042046 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/500c937c-6baf-47e7-93ec-fbff3d5b6d37-bundle" (OuterVolumeSpecName: "bundle") pod "500c937c-6baf-47e7-93ec-fbff3d5b6d37" (UID: "500c937c-6baf-47e7-93ec-fbff3d5b6d37"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:48:59 crc kubenswrapper[4739]: I1201 15:48:59.052782 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/500c937c-6baf-47e7-93ec-fbff3d5b6d37-kube-api-access-22g24" (OuterVolumeSpecName: "kube-api-access-22g24") pod "500c937c-6baf-47e7-93ec-fbff3d5b6d37" (UID: "500c937c-6baf-47e7-93ec-fbff3d5b6d37"). InnerVolumeSpecName "kube-api-access-22g24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:48:59 crc kubenswrapper[4739]: I1201 15:48:59.067586 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/500c937c-6baf-47e7-93ec-fbff3d5b6d37-util" (OuterVolumeSpecName: "util") pod "500c937c-6baf-47e7-93ec-fbff3d5b6d37" (UID: "500c937c-6baf-47e7-93ec-fbff3d5b6d37"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:48:59 crc kubenswrapper[4739]: I1201 15:48:59.142515 4739 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/500c937c-6baf-47e7-93ec-fbff3d5b6d37-util\") on node \"crc\" DevicePath \"\"" Dec 01 15:48:59 crc kubenswrapper[4739]: I1201 15:48:59.142583 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22g24\" (UniqueName: \"kubernetes.io/projected/500c937c-6baf-47e7-93ec-fbff3d5b6d37-kube-api-access-22g24\") on node \"crc\" DevicePath \"\"" Dec 01 15:48:59 crc kubenswrapper[4739]: I1201 15:48:59.142605 4739 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/500c937c-6baf-47e7-93ec-fbff3d5b6d37-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:48:59 crc kubenswrapper[4739]: I1201 15:48:59.575359 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" event={"ID":"500c937c-6baf-47e7-93ec-fbff3d5b6d37","Type":"ContainerDied","Data":"168485495e0c2b882b64308b350d9db53d815a0ac00b708fcb30c92e1f8d365d"} Dec 01 15:48:59 crc kubenswrapper[4739]: I1201 15:48:59.575452 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp" Dec 01 15:48:59 crc kubenswrapper[4739]: I1201 15:48:59.575462 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="168485495e0c2b882b64308b350d9db53d815a0ac00b708fcb30c92e1f8d365d" Dec 01 15:49:05 crc kubenswrapper[4739]: I1201 15:49:05.846357 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-b8d979694-j8gm2"] Dec 01 15:49:05 crc kubenswrapper[4739]: E1201 15:49:05.848238 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="500c937c-6baf-47e7-93ec-fbff3d5b6d37" containerName="util" Dec 01 15:49:05 crc kubenswrapper[4739]: I1201 15:49:05.848272 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="500c937c-6baf-47e7-93ec-fbff3d5b6d37" containerName="util" Dec 01 15:49:05 crc kubenswrapper[4739]: E1201 15:49:05.848312 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="500c937c-6baf-47e7-93ec-fbff3d5b6d37" containerName="pull" Dec 01 15:49:05 crc kubenswrapper[4739]: I1201 15:49:05.848325 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="500c937c-6baf-47e7-93ec-fbff3d5b6d37" containerName="pull" Dec 01 15:49:05 crc kubenswrapper[4739]: E1201 15:49:05.848343 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="500c937c-6baf-47e7-93ec-fbff3d5b6d37" containerName="extract" Dec 01 15:49:05 crc kubenswrapper[4739]: I1201 15:49:05.848357 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="500c937c-6baf-47e7-93ec-fbff3d5b6d37" containerName="extract" Dec 01 15:49:05 crc kubenswrapper[4739]: I1201 15:49:05.848767 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="500c937c-6baf-47e7-93ec-fbff3d5b6d37" containerName="extract" Dec 01 15:49:05 crc kubenswrapper[4739]: I1201 15:49:05.849639 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-b8d979694-j8gm2" Dec 01 15:49:05 crc kubenswrapper[4739]: I1201 15:49:05.851672 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-hfh74" Dec 01 15:49:05 crc kubenswrapper[4739]: I1201 15:49:05.875796 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-b8d979694-j8gm2"] Dec 01 15:49:05 crc kubenswrapper[4739]: I1201 15:49:05.937865 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9swx4\" (UniqueName: \"kubernetes.io/projected/ee802613-91de-4f5b-b0c8-7966498b954c-kube-api-access-9swx4\") pod \"openstack-operator-controller-operator-b8d979694-j8gm2\" (UID: \"ee802613-91de-4f5b-b0c8-7966498b954c\") " pod="openstack-operators/openstack-operator-controller-operator-b8d979694-j8gm2" Dec 01 15:49:06 crc kubenswrapper[4739]: I1201 15:49:06.038997 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9swx4\" (UniqueName: \"kubernetes.io/projected/ee802613-91de-4f5b-b0c8-7966498b954c-kube-api-access-9swx4\") pod \"openstack-operator-controller-operator-b8d979694-j8gm2\" (UID: \"ee802613-91de-4f5b-b0c8-7966498b954c\") " pod="openstack-operators/openstack-operator-controller-operator-b8d979694-j8gm2" Dec 01 15:49:06 crc kubenswrapper[4739]: I1201 15:49:06.062207 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9swx4\" (UniqueName: \"kubernetes.io/projected/ee802613-91de-4f5b-b0c8-7966498b954c-kube-api-access-9swx4\") pod \"openstack-operator-controller-operator-b8d979694-j8gm2\" (UID: \"ee802613-91de-4f5b-b0c8-7966498b954c\") " pod="openstack-operators/openstack-operator-controller-operator-b8d979694-j8gm2" Dec 01 15:49:06 crc kubenswrapper[4739]: I1201 15:49:06.175635 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-b8d979694-j8gm2" Dec 01 15:49:06 crc kubenswrapper[4739]: I1201 15:49:06.412954 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-b8d979694-j8gm2"] Dec 01 15:49:06 crc kubenswrapper[4739]: I1201 15:49:06.626352 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-b8d979694-j8gm2" event={"ID":"ee802613-91de-4f5b-b0c8-7966498b954c","Type":"ContainerStarted","Data":"61a46196ec3cb05fee39cea0e656525219f2429cabe6879f7d725bed826d8536"} Dec 01 15:49:10 crc kubenswrapper[4739]: I1201 15:49:10.648806 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-b8d979694-j8gm2" event={"ID":"ee802613-91de-4f5b-b0c8-7966498b954c","Type":"ContainerStarted","Data":"04aaaec5b34e3e5018a30ee97476980774f27f516fc0cb63a6c571a7da8e0ee5"} Dec 01 15:49:10 crc kubenswrapper[4739]: I1201 15:49:10.649373 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-b8d979694-j8gm2" Dec 01 15:49:10 crc kubenswrapper[4739]: I1201 15:49:10.728003 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-b8d979694-j8gm2" podStartSLOduration=1.7489380350000001 podStartE2EDuration="5.727983584s" podCreationTimestamp="2025-12-01 15:49:05 +0000 UTC" firstStartedPulling="2025-12-01 15:49:06.420499894 +0000 UTC m=+848.246245988" lastFinishedPulling="2025-12-01 15:49:10.399545443 +0000 UTC m=+852.225291537" observedRunningTime="2025-12-01 15:49:10.723441368 +0000 UTC m=+852.549187472" watchObservedRunningTime="2025-12-01 15:49:10.727983584 +0000 UTC m=+852.553729678" Dec 01 15:49:16 crc kubenswrapper[4739]: I1201 15:49:16.178514 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-b8d979694-j8gm2" Dec 01 15:49:16 crc kubenswrapper[4739]: I1201 15:49:16.597247 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tdpjl"] Dec 01 15:49:16 crc kubenswrapper[4739]: I1201 15:49:16.599377 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:16 crc kubenswrapper[4739]: I1201 15:49:16.620677 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tdpjl"] Dec 01 15:49:16 crc kubenswrapper[4739]: I1201 15:49:16.644115 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd062da8-b46d-4c62-849a-43f5039c0b5e-utilities\") pod \"community-operators-tdpjl\" (UID: \"fd062da8-b46d-4c62-849a-43f5039c0b5e\") " pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:16 crc kubenswrapper[4739]: I1201 15:49:16.644197 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x979\" (UniqueName: \"kubernetes.io/projected/fd062da8-b46d-4c62-849a-43f5039c0b5e-kube-api-access-9x979\") pod \"community-operators-tdpjl\" (UID: \"fd062da8-b46d-4c62-849a-43f5039c0b5e\") " pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:16 crc kubenswrapper[4739]: I1201 15:49:16.644355 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd062da8-b46d-4c62-849a-43f5039c0b5e-catalog-content\") pod \"community-operators-tdpjl\" (UID: \"fd062da8-b46d-4c62-849a-43f5039c0b5e\") " pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:16 crc kubenswrapper[4739]: I1201 15:49:16.746395 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd062da8-b46d-4c62-849a-43f5039c0b5e-catalog-content\") pod \"community-operators-tdpjl\" (UID: \"fd062da8-b46d-4c62-849a-43f5039c0b5e\") " pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:16 crc kubenswrapper[4739]: I1201 15:49:16.746573 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd062da8-b46d-4c62-849a-43f5039c0b5e-utilities\") pod \"community-operators-tdpjl\" (UID: \"fd062da8-b46d-4c62-849a-43f5039c0b5e\") " pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:16 crc kubenswrapper[4739]: I1201 15:49:16.746635 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x979\" (UniqueName: \"kubernetes.io/projected/fd062da8-b46d-4c62-849a-43f5039c0b5e-kube-api-access-9x979\") pod \"community-operators-tdpjl\" (UID: \"fd062da8-b46d-4c62-849a-43f5039c0b5e\") " pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:16 crc kubenswrapper[4739]: I1201 15:49:16.747755 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd062da8-b46d-4c62-849a-43f5039c0b5e-utilities\") pod \"community-operators-tdpjl\" (UID: \"fd062da8-b46d-4c62-849a-43f5039c0b5e\") " pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:16 crc kubenswrapper[4739]: I1201 15:49:16.747750 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd062da8-b46d-4c62-849a-43f5039c0b5e-catalog-content\") pod \"community-operators-tdpjl\" (UID: \"fd062da8-b46d-4c62-849a-43f5039c0b5e\") " pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:16 crc kubenswrapper[4739]: I1201 15:49:16.788483 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x979\" (UniqueName: \"kubernetes.io/projected/fd062da8-b46d-4c62-849a-43f5039c0b5e-kube-api-access-9x979\") pod \"community-operators-tdpjl\" (UID: \"fd062da8-b46d-4c62-849a-43f5039c0b5e\") " pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:16 crc kubenswrapper[4739]: I1201 15:49:16.927443 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:17 crc kubenswrapper[4739]: I1201 15:49:17.181085 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tdpjl"] Dec 01 15:49:17 crc kubenswrapper[4739]: I1201 15:49:17.750706 4739 generic.go:334] "Generic (PLEG): container finished" podID="fd062da8-b46d-4c62-849a-43f5039c0b5e" containerID="8593f1fa9e9b784a8495f3bdabec3e2806ddaa52b524d0015232b23664ad4499" exitCode=0 Dec 01 15:49:17 crc kubenswrapper[4739]: I1201 15:49:17.750748 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdpjl" event={"ID":"fd062da8-b46d-4c62-849a-43f5039c0b5e","Type":"ContainerDied","Data":"8593f1fa9e9b784a8495f3bdabec3e2806ddaa52b524d0015232b23664ad4499"} Dec 01 15:49:17 crc kubenswrapper[4739]: I1201 15:49:17.750775 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdpjl" event={"ID":"fd062da8-b46d-4c62-849a-43f5039c0b5e","Type":"ContainerStarted","Data":"23a56edbb57b09d70546a5f261ba671f409c192fd14ee787d90616eeeb76b1fd"} Dec 01 15:49:18 crc kubenswrapper[4739]: I1201 15:49:18.762658 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdpjl" event={"ID":"fd062da8-b46d-4c62-849a-43f5039c0b5e","Type":"ContainerStarted","Data":"9426ad45d9e0d4f7acc32f869c9aee473eb6d554e197aac45366ff96785affdc"} Dec 01 15:49:19 crc kubenswrapper[4739]: I1201 15:49:19.775902 4739 generic.go:334] "Generic (PLEG): container finished" podID="fd062da8-b46d-4c62-849a-43f5039c0b5e" containerID="9426ad45d9e0d4f7acc32f869c9aee473eb6d554e197aac45366ff96785affdc" exitCode=0 Dec 01 15:49:19 crc kubenswrapper[4739]: I1201 15:49:19.776851 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdpjl" event={"ID":"fd062da8-b46d-4c62-849a-43f5039c0b5e","Type":"ContainerDied","Data":"9426ad45d9e0d4f7acc32f869c9aee473eb6d554e197aac45366ff96785affdc"} Dec 01 15:49:20 crc kubenswrapper[4739]: I1201 15:49:20.793286 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdpjl" event={"ID":"fd062da8-b46d-4c62-849a-43f5039c0b5e","Type":"ContainerStarted","Data":"1d0d4a01a0a9de2694de675aa9a853c67757c422af9493adf39042085f1df0bf"} Dec 01 15:49:20 crc kubenswrapper[4739]: I1201 15:49:20.820373 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tdpjl" podStartSLOduration=2.326128004 podStartE2EDuration="4.820361376s" podCreationTimestamp="2025-12-01 15:49:16 +0000 UTC" firstStartedPulling="2025-12-01 15:49:17.752876255 +0000 UTC m=+859.578622359" lastFinishedPulling="2025-12-01 15:49:20.247109597 +0000 UTC m=+862.072855731" observedRunningTime="2025-12-01 15:49:20.817393687 +0000 UTC m=+862.643139781" watchObservedRunningTime="2025-12-01 15:49:20.820361376 +0000 UTC m=+862.646107470" Dec 01 15:49:26 crc kubenswrapper[4739]: I1201 15:49:26.927679 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:26 crc kubenswrapper[4739]: I1201 15:49:26.928662 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:26 crc kubenswrapper[4739]: I1201 15:49:26.985588 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:27 crc kubenswrapper[4739]: I1201 15:49:27.933284 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:28 crc kubenswrapper[4739]: I1201 15:49:28.005758 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tdpjl"] Dec 01 15:49:29 crc kubenswrapper[4739]: I1201 15:49:29.868731 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tdpjl" podUID="fd062da8-b46d-4c62-849a-43f5039c0b5e" containerName="registry-server" containerID="cri-o://1d0d4a01a0a9de2694de675aa9a853c67757c422af9493adf39042085f1df0bf" gracePeriod=2 Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.729905 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.754052 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd062da8-b46d-4c62-849a-43f5039c0b5e-utilities\") pod \"fd062da8-b46d-4c62-849a-43f5039c0b5e\" (UID: \"fd062da8-b46d-4c62-849a-43f5039c0b5e\") " Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.754404 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd062da8-b46d-4c62-849a-43f5039c0b5e-catalog-content\") pod \"fd062da8-b46d-4c62-849a-43f5039c0b5e\" (UID: \"fd062da8-b46d-4c62-849a-43f5039c0b5e\") " Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.755000 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd062da8-b46d-4c62-849a-43f5039c0b5e-utilities" (OuterVolumeSpecName: "utilities") pod "fd062da8-b46d-4c62-849a-43f5039c0b5e" (UID: "fd062da8-b46d-4c62-849a-43f5039c0b5e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.755136 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9x979\" (UniqueName: \"kubernetes.io/projected/fd062da8-b46d-4c62-849a-43f5039c0b5e-kube-api-access-9x979\") pod \"fd062da8-b46d-4c62-849a-43f5039c0b5e\" (UID: \"fd062da8-b46d-4c62-849a-43f5039c0b5e\") " Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.755609 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd062da8-b46d-4c62-849a-43f5039c0b5e-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.759666 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd062da8-b46d-4c62-849a-43f5039c0b5e-kube-api-access-9x979" (OuterVolumeSpecName: "kube-api-access-9x979") pod "fd062da8-b46d-4c62-849a-43f5039c0b5e" (UID: "fd062da8-b46d-4c62-849a-43f5039c0b5e"). InnerVolumeSpecName "kube-api-access-9x979". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.814888 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd062da8-b46d-4c62-849a-43f5039c0b5e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd062da8-b46d-4c62-849a-43f5039c0b5e" (UID: "fd062da8-b46d-4c62-849a-43f5039c0b5e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.856505 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9x979\" (UniqueName: \"kubernetes.io/projected/fd062da8-b46d-4c62-849a-43f5039c0b5e-kube-api-access-9x979\") on node \"crc\" DevicePath \"\"" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.856878 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd062da8-b46d-4c62-849a-43f5039c0b5e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.875477 4739 generic.go:334] "Generic (PLEG): container finished" podID="fd062da8-b46d-4c62-849a-43f5039c0b5e" containerID="1d0d4a01a0a9de2694de675aa9a853c67757c422af9493adf39042085f1df0bf" exitCode=0 Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.875516 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdpjl" event={"ID":"fd062da8-b46d-4c62-849a-43f5039c0b5e","Type":"ContainerDied","Data":"1d0d4a01a0a9de2694de675aa9a853c67757c422af9493adf39042085f1df0bf"} Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.875541 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdpjl" event={"ID":"fd062da8-b46d-4c62-849a-43f5039c0b5e","Type":"ContainerDied","Data":"23a56edbb57b09d70546a5f261ba671f409c192fd14ee787d90616eeeb76b1fd"} Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.875557 4739 scope.go:117] "RemoveContainer" containerID="1d0d4a01a0a9de2694de675aa9a853c67757c422af9493adf39042085f1df0bf" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.875576 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdpjl" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.897165 4739 scope.go:117] "RemoveContainer" containerID="9426ad45d9e0d4f7acc32f869c9aee473eb6d554e197aac45366ff96785affdc" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.911206 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tdpjl"] Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.917328 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tdpjl"] Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.940709 4739 scope.go:117] "RemoveContainer" containerID="8593f1fa9e9b784a8495f3bdabec3e2806ddaa52b524d0015232b23664ad4499" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.957145 4739 scope.go:117] "RemoveContainer" containerID="1d0d4a01a0a9de2694de675aa9a853c67757c422af9493adf39042085f1df0bf" Dec 01 15:49:30 crc kubenswrapper[4739]: E1201 15:49:30.957632 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d0d4a01a0a9de2694de675aa9a853c67757c422af9493adf39042085f1df0bf\": container with ID starting with 1d0d4a01a0a9de2694de675aa9a853c67757c422af9493adf39042085f1df0bf not found: ID does not exist" containerID="1d0d4a01a0a9de2694de675aa9a853c67757c422af9493adf39042085f1df0bf" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.957712 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d0d4a01a0a9de2694de675aa9a853c67757c422af9493adf39042085f1df0bf"} err="failed to get container status \"1d0d4a01a0a9de2694de675aa9a853c67757c422af9493adf39042085f1df0bf\": rpc error: code = NotFound desc = could not find container \"1d0d4a01a0a9de2694de675aa9a853c67757c422af9493adf39042085f1df0bf\": container with ID starting with 1d0d4a01a0a9de2694de675aa9a853c67757c422af9493adf39042085f1df0bf not found: ID does not exist" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.957788 4739 scope.go:117] "RemoveContainer" containerID="9426ad45d9e0d4f7acc32f869c9aee473eb6d554e197aac45366ff96785affdc" Dec 01 15:49:30 crc kubenswrapper[4739]: E1201 15:49:30.958384 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9426ad45d9e0d4f7acc32f869c9aee473eb6d554e197aac45366ff96785affdc\": container with ID starting with 9426ad45d9e0d4f7acc32f869c9aee473eb6d554e197aac45366ff96785affdc not found: ID does not exist" containerID="9426ad45d9e0d4f7acc32f869c9aee473eb6d554e197aac45366ff96785affdc" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.958552 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9426ad45d9e0d4f7acc32f869c9aee473eb6d554e197aac45366ff96785affdc"} err="failed to get container status \"9426ad45d9e0d4f7acc32f869c9aee473eb6d554e197aac45366ff96785affdc\": rpc error: code = NotFound desc = could not find container \"9426ad45d9e0d4f7acc32f869c9aee473eb6d554e197aac45366ff96785affdc\": container with ID starting with 9426ad45d9e0d4f7acc32f869c9aee473eb6d554e197aac45366ff96785affdc not found: ID does not exist" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.958658 4739 scope.go:117] "RemoveContainer" containerID="8593f1fa9e9b784a8495f3bdabec3e2806ddaa52b524d0015232b23664ad4499" Dec 01 15:49:30 crc kubenswrapper[4739]: E1201 15:49:30.959085 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8593f1fa9e9b784a8495f3bdabec3e2806ddaa52b524d0015232b23664ad4499\": container with ID starting with 8593f1fa9e9b784a8495f3bdabec3e2806ddaa52b524d0015232b23664ad4499 not found: ID does not exist" containerID="8593f1fa9e9b784a8495f3bdabec3e2806ddaa52b524d0015232b23664ad4499" Dec 01 15:49:30 crc kubenswrapper[4739]: I1201 15:49:30.959167 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8593f1fa9e9b784a8495f3bdabec3e2806ddaa52b524d0015232b23664ad4499"} err="failed to get container status \"8593f1fa9e9b784a8495f3bdabec3e2806ddaa52b524d0015232b23664ad4499\": rpc error: code = NotFound desc = could not find container \"8593f1fa9e9b784a8495f3bdabec3e2806ddaa52b524d0015232b23664ad4499\": container with ID starting with 8593f1fa9e9b784a8495f3bdabec3e2806ddaa52b524d0015232b23664ad4499 not found: ID does not exist" Dec 01 15:49:30 crc kubenswrapper[4739]: E1201 15:49:30.976497 4739 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd062da8_b46d_4c62_849a_43f5039c0b5e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd062da8_b46d_4c62_849a_43f5039c0b5e.slice/crio-23a56edbb57b09d70546a5f261ba671f409c192fd14ee787d90616eeeb76b1fd\": RecentStats: unable to find data in memory cache]" Dec 01 15:49:32 crc kubenswrapper[4739]: I1201 15:49:32.485793 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd062da8-b46d-4c62-849a-43f5039c0b5e" path="/var/lib/kubelet/pods/fd062da8-b46d-4c62-849a-43f5039c0b5e/volumes" Dec 01 15:49:39 crc kubenswrapper[4739]: I1201 15:49:39.622641 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:49:39 crc kubenswrapper[4739]: I1201 15:49:39.624490 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.394520 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-nz6dz"] Dec 01 15:49:49 crc kubenswrapper[4739]: E1201 15:49:49.395320 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd062da8-b46d-4c62-849a-43f5039c0b5e" containerName="extract-content" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.395335 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd062da8-b46d-4c62-849a-43f5039c0b5e" containerName="extract-content" Dec 01 15:49:49 crc kubenswrapper[4739]: E1201 15:49:49.395351 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd062da8-b46d-4c62-849a-43f5039c0b5e" containerName="extract-utilities" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.395387 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd062da8-b46d-4c62-849a-43f5039c0b5e" containerName="extract-utilities" Dec 01 15:49:49 crc kubenswrapper[4739]: E1201 15:49:49.395405 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd062da8-b46d-4c62-849a-43f5039c0b5e" containerName="registry-server" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.395416 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd062da8-b46d-4c62-849a-43f5039c0b5e" containerName="registry-server" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.395609 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd062da8-b46d-4c62-849a-43f5039c0b5e" containerName="registry-server" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.396328 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nz6dz" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.402277 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.403444 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.412776 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-nz6dz"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.422013 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-cg6lv" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.422301 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-8gspt" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.431350 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvg2l\" (UniqueName: \"kubernetes.io/projected/97f978b1-c0ba-4fa7-8a8f-91de3ab80bc5-kube-api-access-bvg2l\") pod \"barbican-operator-controller-manager-7d9dfd778-nz6dz\" (UID: \"97f978b1-c0ba-4fa7-8a8f-91de3ab80bc5\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nz6dz" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.431486 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdht5\" (UniqueName: \"kubernetes.io/projected/28db1029-7abe-4d58-afbd-0022422c286c-kube-api-access-tdht5\") pod \"cinder-operator-controller-manager-5d986d85c9-2s7sq\" (UID: \"28db1029-7abe-4d58-afbd-0022422c286c\") " pod="openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.435804 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.458786 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-th7k2"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.460784 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-th7k2" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.465233 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-q98rd" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.499370 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-2m26g"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.501200 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-2m26g" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.502539 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-th7k2"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.510200 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-rcl56" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.516860 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-2m26g"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.523715 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmntr"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.525945 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmntr" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.528092 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-9rrv6" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.532633 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdht5\" (UniqueName: \"kubernetes.io/projected/28db1029-7abe-4d58-afbd-0022422c286c-kube-api-access-tdht5\") pod \"cinder-operator-controller-manager-5d986d85c9-2s7sq\" (UID: \"28db1029-7abe-4d58-afbd-0022422c286c\") " pod="openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.532706 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvg2l\" (UniqueName: \"kubernetes.io/projected/97f978b1-c0ba-4fa7-8a8f-91de3ab80bc5-kube-api-access-bvg2l\") pod \"barbican-operator-controller-manager-7d9dfd778-nz6dz\" (UID: \"97f978b1-c0ba-4fa7-8a8f-91de3ab80bc5\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nz6dz" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.536071 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmntr"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.561599 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-f2t8s"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.562612 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-f2t8s" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.564577 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-zkbl8" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.569512 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.570695 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-f2t8s"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.570796 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.577715 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvg2l\" (UniqueName: \"kubernetes.io/projected/97f978b1-c0ba-4fa7-8a8f-91de3ab80bc5-kube-api-access-bvg2l\") pod \"barbican-operator-controller-manager-7d9dfd778-nz6dz\" (UID: \"97f978b1-c0ba-4fa7-8a8f-91de3ab80bc5\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nz6dz" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.578665 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdht5\" (UniqueName: \"kubernetes.io/projected/28db1029-7abe-4d58-afbd-0022422c286c-kube-api-access-tdht5\") pod \"cinder-operator-controller-manager-5d986d85c9-2s7sq\" (UID: \"28db1029-7abe-4d58-afbd-0022422c286c\") " pod="openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.585540 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.585783 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-gwc9m" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.587556 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-pqqzm"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.588556 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-pqqzm" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.590769 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-s9fcv" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.604436 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.621565 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-pqqzm"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.636145 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slm67\" (UniqueName: \"kubernetes.io/projected/895e540c-7e60-45b1-8275-36588bb8e788-kube-api-access-slm67\") pod \"horizon-operator-controller-manager-68c6d99b8f-f2t8s\" (UID: \"895e540c-7e60-45b1-8275-36588bb8e788\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-f2t8s" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.636211 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccht4\" (UniqueName: \"kubernetes.io/projected/8525c2bf-aedc-4816-8ea5-4ebdda160f3d-kube-api-access-ccht4\") pod \"glance-operator-controller-manager-668d9c48b9-2m26g\" (UID: \"8525c2bf-aedc-4816-8ea5-4ebdda160f3d\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-2m26g" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.636729 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrplg\" (UniqueName: \"kubernetes.io/projected/44a9534e-febc-4a41-9ee0-af96baa95a78-kube-api-access-zrplg\") pod \"designate-operator-controller-manager-78b4bc895b-th7k2\" (UID: \"44a9534e-febc-4a41-9ee0-af96baa95a78\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-th7k2" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.636768 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hccs\" (UniqueName: \"kubernetes.io/projected/3422b21c-7fef-4c3a-8990-85fa0d40b027-kube-api-access-6hccs\") pod \"heat-operator-controller-manager-5f64f6f8bb-tmntr\" (UID: \"3422b21c-7fef-4c3a-8990-85fa0d40b027\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmntr" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.637183 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert\") pod \"infra-operator-controller-manager-57548d458d-gpkcn\" (UID: \"54dcd8b3-40f6-4487-bc4f-810c6aebb6fe\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.637215 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrrqf\" (UniqueName: \"kubernetes.io/projected/ebe9e6cb-7cc7-4a43-ba59-ff0b8209d314-kube-api-access-hrrqf\") pod \"ironic-operator-controller-manager-6c548fd776-pqqzm\" (UID: \"ebe9e6cb-7cc7-4a43-ba59-ff0b8209d314\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-pqqzm" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.637262 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2vzg\" (UniqueName: \"kubernetes.io/projected/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-kube-api-access-j2vzg\") pod \"infra-operator-controller-manager-57548d458d-gpkcn\" (UID: \"54dcd8b3-40f6-4487-bc4f-810c6aebb6fe\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.642372 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.643327 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.649871 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-95srg" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.673537 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.702095 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-d2dxq"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.703008 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-d2dxq" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.704492 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-d2dxq"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.704725 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-khr2h" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.711642 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g7bc7"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.712735 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g7bc7" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.714748 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-vjl64" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.737799 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2vzg\" (UniqueName: \"kubernetes.io/projected/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-kube-api-access-j2vzg\") pod \"infra-operator-controller-manager-57548d458d-gpkcn\" (UID: \"54dcd8b3-40f6-4487-bc4f-810c6aebb6fe\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.737869 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slm67\" (UniqueName: \"kubernetes.io/projected/895e540c-7e60-45b1-8275-36588bb8e788-kube-api-access-slm67\") pod \"horizon-operator-controller-manager-68c6d99b8f-f2t8s\" (UID: \"895e540c-7e60-45b1-8275-36588bb8e788\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-f2t8s" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.737901 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccht4\" (UniqueName: \"kubernetes.io/projected/8525c2bf-aedc-4816-8ea5-4ebdda160f3d-kube-api-access-ccht4\") pod \"glance-operator-controller-manager-668d9c48b9-2m26g\" (UID: \"8525c2bf-aedc-4816-8ea5-4ebdda160f3d\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-2m26g" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.737926 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4nzr\" (UniqueName: \"kubernetes.io/projected/644da74f-c26b-4fae-a81f-4091a1e95b87-kube-api-access-m4nzr\") pod \"mariadb-operator-controller-manager-56bbcc9d85-g7bc7\" (UID: \"644da74f-c26b-4fae-a81f-4091a1e95b87\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g7bc7" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.737968 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrplg\" (UniqueName: \"kubernetes.io/projected/44a9534e-febc-4a41-9ee0-af96baa95a78-kube-api-access-zrplg\") pod \"designate-operator-controller-manager-78b4bc895b-th7k2\" (UID: \"44a9534e-febc-4a41-9ee0-af96baa95a78\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-th7k2" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.737993 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hccs\" (UniqueName: \"kubernetes.io/projected/3422b21c-7fef-4c3a-8990-85fa0d40b027-kube-api-access-6hccs\") pod \"heat-operator-controller-manager-5f64f6f8bb-tmntr\" (UID: \"3422b21c-7fef-4c3a-8990-85fa0d40b027\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmntr" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.738021 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cdnw\" (UniqueName: \"kubernetes.io/projected/8f8c53db-885f-4548-bcc0-02ce36cd6543-kube-api-access-6cdnw\") pod \"keystone-operator-controller-manager-546d4bdf48-758wd\" (UID: \"8f8c53db-885f-4548-bcc0-02ce36cd6543\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.738067 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert\") pod \"infra-operator-controller-manager-57548d458d-gpkcn\" (UID: \"54dcd8b3-40f6-4487-bc4f-810c6aebb6fe\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.738094 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrrqf\" (UniqueName: \"kubernetes.io/projected/ebe9e6cb-7cc7-4a43-ba59-ff0b8209d314-kube-api-access-hrrqf\") pod \"ironic-operator-controller-manager-6c548fd776-pqqzm\" (UID: \"ebe9e6cb-7cc7-4a43-ba59-ff0b8209d314\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-pqqzm" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.738126 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngmp2\" (UniqueName: \"kubernetes.io/projected/7a27f279-7c57-439c-88f2-67914f2e7b9e-kube-api-access-ngmp2\") pod \"manila-operator-controller-manager-6546668bfd-d2dxq\" (UID: \"7a27f279-7c57-439c-88f2-67914f2e7b9e\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-d2dxq" Dec 01 15:49:49 crc kubenswrapper[4739]: E1201 15:49:49.739124 4739 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 15:49:49 crc kubenswrapper[4739]: E1201 15:49:49.739234 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert podName:54dcd8b3-40f6-4487-bc4f-810c6aebb6fe nodeName:}" failed. No retries permitted until 2025-12-01 15:49:50.239218127 +0000 UTC m=+892.064964211 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert") pod "infra-operator-controller-manager-57548d458d-gpkcn" (UID: "54dcd8b3-40f6-4487-bc4f-810c6aebb6fe") : secret "infra-operator-webhook-server-cert" not found Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.744486 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g7bc7"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.745007 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nz6dz" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.753134 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jrc6n"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.755878 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jrc6n" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.760171 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-rdnwt" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.761392 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.768005 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.768976 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hccs\" (UniqueName: \"kubernetes.io/projected/3422b21c-7fef-4c3a-8990-85fa0d40b027-kube-api-access-6hccs\") pod \"heat-operator-controller-manager-5f64f6f8bb-tmntr\" (UID: \"3422b21c-7fef-4c3a-8990-85fa0d40b027\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmntr" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.769529 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.772855 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2vzg\" (UniqueName: \"kubernetes.io/projected/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-kube-api-access-j2vzg\") pod \"infra-operator-controller-manager-57548d458d-gpkcn\" (UID: \"54dcd8b3-40f6-4487-bc4f-810c6aebb6fe\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.773185 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slm67\" (UniqueName: \"kubernetes.io/projected/895e540c-7e60-45b1-8275-36588bb8e788-kube-api-access-slm67\") pod \"horizon-operator-controller-manager-68c6d99b8f-f2t8s\" (UID: \"895e540c-7e60-45b1-8275-36588bb8e788\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-f2t8s" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.774112 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrrqf\" (UniqueName: \"kubernetes.io/projected/ebe9e6cb-7cc7-4a43-ba59-ff0b8209d314-kube-api-access-hrrqf\") pod \"ironic-operator-controller-manager-6c548fd776-pqqzm\" (UID: \"ebe9e6cb-7cc7-4a43-ba59-ff0b8209d314\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-pqqzm" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.778895 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-w7v2q" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.780558 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jrc6n"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.783578 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.784713 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrplg\" (UniqueName: \"kubernetes.io/projected/44a9534e-febc-4a41-9ee0-af96baa95a78-kube-api-access-zrplg\") pod \"designate-operator-controller-manager-78b4bc895b-th7k2\" (UID: \"44a9534e-febc-4a41-9ee0-af96baa95a78\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-th7k2" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.801026 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-kkchm"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.805635 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-th7k2" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.806812 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-kkchm"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.806888 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkchm" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.813532 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.815068 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.818971 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-dwn29" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.819493 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.819546 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.824667 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-9fspg" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.827643 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-4nwgp"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.829149 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccht4\" (UniqueName: \"kubernetes.io/projected/8525c2bf-aedc-4816-8ea5-4ebdda160f3d-kube-api-access-ccht4\") pod \"glance-operator-controller-manager-668d9c48b9-2m26g\" (UID: \"8525c2bf-aedc-4816-8ea5-4ebdda160f3d\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-2m26g" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.829157 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4nwgp" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.837685 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-zrhx2"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.839145 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-nphbk" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.840471 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zrhx2" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.843058 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-xqc94" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.843691 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngmp2\" (UniqueName: \"kubernetes.io/projected/7a27f279-7c57-439c-88f2-67914f2e7b9e-kube-api-access-ngmp2\") pod \"manila-operator-controller-manager-6546668bfd-d2dxq\" (UID: \"7a27f279-7c57-439c-88f2-67914f2e7b9e\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-d2dxq" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.843795 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4nzr\" (UniqueName: \"kubernetes.io/projected/644da74f-c26b-4fae-a81f-4091a1e95b87-kube-api-access-m4nzr\") pod \"mariadb-operator-controller-manager-56bbcc9d85-g7bc7\" (UID: \"644da74f-c26b-4fae-a81f-4091a1e95b87\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g7bc7" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.843874 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cdnw\" (UniqueName: \"kubernetes.io/projected/8f8c53db-885f-4548-bcc0-02ce36cd6543-kube-api-access-6cdnw\") pod \"keystone-operator-controller-manager-546d4bdf48-758wd\" (UID: \"8f8c53db-885f-4548-bcc0-02ce36cd6543\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.848410 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmntr" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.897072 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cdnw\" (UniqueName: \"kubernetes.io/projected/8f8c53db-885f-4548-bcc0-02ce36cd6543-kube-api-access-6cdnw\") pod \"keystone-operator-controller-manager-546d4bdf48-758wd\" (UID: \"8f8c53db-885f-4548-bcc0-02ce36cd6543\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.897402 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4nzr\" (UniqueName: \"kubernetes.io/projected/644da74f-c26b-4fae-a81f-4091a1e95b87-kube-api-access-m4nzr\") pod \"mariadb-operator-controller-manager-56bbcc9d85-g7bc7\" (UID: \"644da74f-c26b-4fae-a81f-4091a1e95b87\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g7bc7" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.897634 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-zrhx2"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.908185 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngmp2\" (UniqueName: \"kubernetes.io/projected/7a27f279-7c57-439c-88f2-67914f2e7b9e-kube-api-access-ngmp2\") pod \"manila-operator-controller-manager-6546668bfd-d2dxq\" (UID: \"7a27f279-7c57-439c-88f2-67914f2e7b9e\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-d2dxq" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.916215 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-4nwgp"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.922536 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.923626 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.925083 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-bdqnd" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.933804 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-f2t8s" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.945614 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9227w\" (UniqueName: \"kubernetes.io/projected/a4706184-6b4f-4416-b9d6-430087d5577c-kube-api-access-9227w\") pod \"octavia-operator-controller-manager-998648c74-kkchm\" (UID: \"a4706184-6b4f-4416-b9d6-430087d5577c\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkchm" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.946731 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b26mh\" (UniqueName: \"kubernetes.io/projected/903f2159-9ae0-4b78-8447-3b5bd4a420ef-kube-api-access-b26mh\") pod \"ovn-operator-controller-manager-b6456fdb6-4nwgp\" (UID: \"903f2159-9ae0-4b78-8447-3b5bd4a420ef\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4nwgp" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.946874 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsz6x\" (UniqueName: \"kubernetes.io/projected/c6ddb146-59dc-41c2-9278-466eda24296c-kube-api-access-jsz6x\") pod \"placement-operator-controller-manager-78f8948974-zrhx2\" (UID: \"c6ddb146-59dc-41c2-9278-466eda24296c\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-zrhx2" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.946993 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk8jk\" (UniqueName: \"kubernetes.io/projected/974d1918-6e6a-4660-898c-b8d8db439f46-kube-api-access-xk8jk\") pod \"nova-operator-controller-manager-697bc559fc-8bsmt\" (UID: \"974d1918-6e6a-4660-898c-b8d8db439f46\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.947105 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn4hs\" (UniqueName: \"kubernetes.io/projected/5d6c8b0a-e856-4794-a3ce-f72e1f6d0c91-kube-api-access-pn4hs\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-jrc6n\" (UID: \"5d6c8b0a-e856-4794-a3ce-f72e1f6d0c91\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jrc6n" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.947207 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krz24\" (UniqueName: \"kubernetes.io/projected/81c50b9d-cef2-40b0-bb96-0e44676eaf93-kube-api-access-krz24\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh\" (UID: \"81c50b9d-cef2-40b0-bb96-0e44676eaf93\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.947311 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh\" (UID: \"81c50b9d-cef2-40b0-bb96-0e44676eaf93\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.967369 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-pqqzm" Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.972077 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt"] Dec 01 15:49:49 crc kubenswrapper[4739]: I1201 15:49:49.982787 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.000803 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.001938 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.011646 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-kc8cf" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.022086 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.029079 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-d2dxq" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.038331 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g7bc7" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.048993 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk8jk\" (UniqueName: \"kubernetes.io/projected/974d1918-6e6a-4660-898c-b8d8db439f46-kube-api-access-xk8jk\") pod \"nova-operator-controller-manager-697bc559fc-8bsmt\" (UID: \"974d1918-6e6a-4660-898c-b8d8db439f46\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.049876 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn4hs\" (UniqueName: \"kubernetes.io/projected/5d6c8b0a-e856-4794-a3ce-f72e1f6d0c91-kube-api-access-pn4hs\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-jrc6n\" (UID: \"5d6c8b0a-e856-4794-a3ce-f72e1f6d0c91\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jrc6n" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.049904 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krz24\" (UniqueName: \"kubernetes.io/projected/81c50b9d-cef2-40b0-bb96-0e44676eaf93-kube-api-access-krz24\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh\" (UID: \"81c50b9d-cef2-40b0-bb96-0e44676eaf93\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.049939 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh\" (UID: \"81c50b9d-cef2-40b0-bb96-0e44676eaf93\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.049983 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9227w\" (UniqueName: \"kubernetes.io/projected/a4706184-6b4f-4416-b9d6-430087d5577c-kube-api-access-9227w\") pod \"octavia-operator-controller-manager-998648c74-kkchm\" (UID: \"a4706184-6b4f-4416-b9d6-430087d5577c\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkchm" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.050023 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b26mh\" (UniqueName: \"kubernetes.io/projected/903f2159-9ae0-4b78-8447-3b5bd4a420ef-kube-api-access-b26mh\") pod \"ovn-operator-controller-manager-b6456fdb6-4nwgp\" (UID: \"903f2159-9ae0-4b78-8447-3b5bd4a420ef\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4nwgp" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.050044 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g49c\" (UniqueName: \"kubernetes.io/projected/7ff65da3-52d5-4e80-9f48-d23d3bc22ff6-kube-api-access-9g49c\") pod \"swift-operator-controller-manager-5f8c65bbfc-xgpkt\" (UID: \"7ff65da3-52d5-4e80-9f48-d23d3bc22ff6\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.050071 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsz6x\" (UniqueName: \"kubernetes.io/projected/c6ddb146-59dc-41c2-9278-466eda24296c-kube-api-access-jsz6x\") pod \"placement-operator-controller-manager-78f8948974-zrhx2\" (UID: \"c6ddb146-59dc-41c2-9278-466eda24296c\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-zrhx2" Dec 01 15:49:50 crc kubenswrapper[4739]: E1201 15:49:50.051108 4739 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 15:49:50 crc kubenswrapper[4739]: E1201 15:49:50.051143 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert podName:81c50b9d-cef2-40b0-bb96-0e44676eaf93 nodeName:}" failed. No retries permitted until 2025-12-01 15:49:50.551131772 +0000 UTC m=+892.376877856 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" (UID: "81c50b9d-cef2-40b0-bb96-0e44676eaf93") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.073184 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.074255 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.078348 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b26mh\" (UniqueName: \"kubernetes.io/projected/903f2159-9ae0-4b78-8447-3b5bd4a420ef-kube-api-access-b26mh\") pod \"ovn-operator-controller-manager-b6456fdb6-4nwgp\" (UID: \"903f2159-9ae0-4b78-8447-3b5bd4a420ef\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4nwgp" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.080839 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-wh8sh" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.081534 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn4hs\" (UniqueName: \"kubernetes.io/projected/5d6c8b0a-e856-4794-a3ce-f72e1f6d0c91-kube-api-access-pn4hs\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-jrc6n\" (UID: \"5d6c8b0a-e856-4794-a3ce-f72e1f6d0c91\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jrc6n" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.081643 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk8jk\" (UniqueName: \"kubernetes.io/projected/974d1918-6e6a-4660-898c-b8d8db439f46-kube-api-access-xk8jk\") pod \"nova-operator-controller-manager-697bc559fc-8bsmt\" (UID: \"974d1918-6e6a-4660-898c-b8d8db439f46\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.084562 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9227w\" (UniqueName: \"kubernetes.io/projected/a4706184-6b4f-4416-b9d6-430087d5577c-kube-api-access-9227w\") pod \"octavia-operator-controller-manager-998648c74-kkchm\" (UID: \"a4706184-6b4f-4416-b9d6-430087d5577c\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkchm" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.085595 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.089386 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsz6x\" (UniqueName: \"kubernetes.io/projected/c6ddb146-59dc-41c2-9278-466eda24296c-kube-api-access-jsz6x\") pod \"placement-operator-controller-manager-78f8948974-zrhx2\" (UID: \"c6ddb146-59dc-41c2-9278-466eda24296c\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-zrhx2" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.093827 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krz24\" (UniqueName: \"kubernetes.io/projected/81c50b9d-cef2-40b0-bb96-0e44676eaf93-kube-api-access-krz24\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh\" (UID: \"81c50b9d-cef2-40b0-bb96-0e44676eaf93\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.100921 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-px5zx"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.102196 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-px5zx" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.104407 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-cxq9n" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.117234 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-px5zx"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.130779 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-2m26g" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.152790 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cmll\" (UniqueName: \"kubernetes.io/projected/84c3dbfb-b9ac-4b9d-b942-e454b50bf132-kube-api-access-8cmll\") pod \"watcher-operator-controller-manager-769dc69bc-px5zx\" (UID: \"84c3dbfb-b9ac-4b9d-b942-e454b50bf132\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-px5zx" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.152839 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g49c\" (UniqueName: \"kubernetes.io/projected/7ff65da3-52d5-4e80-9f48-d23d3bc22ff6-kube-api-access-9g49c\") pod \"swift-operator-controller-manager-5f8c65bbfc-xgpkt\" (UID: \"7ff65da3-52d5-4e80-9f48-d23d3bc22ff6\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.152889 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54nx7\" (UniqueName: \"kubernetes.io/projected/f60761e0-1afe-46fa-90fc-5ed9f42f730d-kube-api-access-54nx7\") pod \"telemetry-operator-controller-manager-76cc84c6bb-fxvzt\" (UID: \"f60761e0-1afe-46fa-90fc-5ed9f42f730d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.152910 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm8qg\" (UniqueName: \"kubernetes.io/projected/bed989ab-cd94-41f1-91ea-8c1bdfc056af-kube-api-access-lm8qg\") pod \"test-operator-controller-manager-5854674fcc-4pwqm\" (UID: \"bed989ab-cd94-41f1-91ea-8c1bdfc056af\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.174710 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g49c\" (UniqueName: \"kubernetes.io/projected/7ff65da3-52d5-4e80-9f48-d23d3bc22ff6-kube-api-access-9g49c\") pod \"swift-operator-controller-manager-5f8c65bbfc-xgpkt\" (UID: \"7ff65da3-52d5-4e80-9f48-d23d3bc22ff6\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.180799 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.181765 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.184163 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.185821 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.185958 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.186196 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-qscx2" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.199809 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jrc6n" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.242972 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qm8fw"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.244992 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.246026 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qm8fw" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.249403 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-nwc7g" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.257338 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cmll\" (UniqueName: \"kubernetes.io/projected/84c3dbfb-b9ac-4b9d-b942-e454b50bf132-kube-api-access-8cmll\") pod \"watcher-operator-controller-manager-769dc69bc-px5zx\" (UID: \"84c3dbfb-b9ac-4b9d-b942-e454b50bf132\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-px5zx" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.257408 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54nx7\" (UniqueName: \"kubernetes.io/projected/f60761e0-1afe-46fa-90fc-5ed9f42f730d-kube-api-access-54nx7\") pod \"telemetry-operator-controller-manager-76cc84c6bb-fxvzt\" (UID: \"f60761e0-1afe-46fa-90fc-5ed9f42f730d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.257446 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm8qg\" (UniqueName: \"kubernetes.io/projected/bed989ab-cd94-41f1-91ea-8c1bdfc056af-kube-api-access-lm8qg\") pod \"test-operator-controller-manager-5854674fcc-4pwqm\" (UID: \"bed989ab-cd94-41f1-91ea-8c1bdfc056af\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.257481 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert\") pod \"infra-operator-controller-manager-57548d458d-gpkcn\" (UID: \"54dcd8b3-40f6-4487-bc4f-810c6aebb6fe\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:49:50 crc kubenswrapper[4739]: E1201 15:49:50.257607 4739 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 15:49:50 crc kubenswrapper[4739]: E1201 15:49:50.257646 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert podName:54dcd8b3-40f6-4487-bc4f-810c6aebb6fe nodeName:}" failed. No retries permitted until 2025-12-01 15:49:51.257633268 +0000 UTC m=+893.083379362 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert") pod "infra-operator-controller-manager-57548d458d-gpkcn" (UID: "54dcd8b3-40f6-4487-bc4f-810c6aebb6fe") : secret "infra-operator-webhook-server-cert" not found Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.260764 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkchm" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.284154 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm8qg\" (UniqueName: \"kubernetes.io/projected/bed989ab-cd94-41f1-91ea-8c1bdfc056af-kube-api-access-lm8qg\") pod \"test-operator-controller-manager-5854674fcc-4pwqm\" (UID: \"bed989ab-cd94-41f1-91ea-8c1bdfc056af\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.293356 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54nx7\" (UniqueName: \"kubernetes.io/projected/f60761e0-1afe-46fa-90fc-5ed9f42f730d-kube-api-access-54nx7\") pod \"telemetry-operator-controller-manager-76cc84c6bb-fxvzt\" (UID: \"f60761e0-1afe-46fa-90fc-5ed9f42f730d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.296842 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cmll\" (UniqueName: \"kubernetes.io/projected/84c3dbfb-b9ac-4b9d-b942-e454b50bf132-kube-api-access-8cmll\") pod \"watcher-operator-controller-manager-769dc69bc-px5zx\" (UID: \"84c3dbfb-b9ac-4b9d-b942-e454b50bf132\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-px5zx" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.296909 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qm8fw"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.319333 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4nwgp" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.336226 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zrhx2" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.358717 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.359081 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcmbv\" (UniqueName: \"kubernetes.io/projected/5f474df5-ced8-45ff-840a-c52898696bf7-kube-api-access-pcmbv\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.359134 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.359161 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m969w\" (UniqueName: \"kubernetes.io/projected/47bf6f4a-97c7-4f1f-a00d-0c5151058f2c-kube-api-access-m969w\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qm8fw\" (UID: \"47bf6f4a-97c7-4f1f-a00d-0c5151058f2c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qm8fw" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.359217 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.380833 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-nz6dz"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.395651 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.414231 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.419456 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-th7k2"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.432988 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-px5zx" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.490530 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcmbv\" (UniqueName: \"kubernetes.io/projected/5f474df5-ced8-45ff-840a-c52898696bf7-kube-api-access-pcmbv\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.490574 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.490608 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m969w\" (UniqueName: \"kubernetes.io/projected/47bf6f4a-97c7-4f1f-a00d-0c5151058f2c-kube-api-access-m969w\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qm8fw\" (UID: \"47bf6f4a-97c7-4f1f-a00d-0c5151058f2c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qm8fw" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.490655 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:50 crc kubenswrapper[4739]: E1201 15:49:50.490790 4739 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 15:49:50 crc kubenswrapper[4739]: E1201 15:49:50.490834 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs podName:5f474df5-ced8-45ff-840a-c52898696bf7 nodeName:}" failed. No retries permitted until 2025-12-01 15:49:50.990821057 +0000 UTC m=+892.816567151 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs") pod "openstack-operator-controller-manager-59fc5775d4-chxkc" (UID: "5f474df5-ced8-45ff-840a-c52898696bf7") : secret "webhook-server-cert" not found Dec 01 15:49:50 crc kubenswrapper[4739]: E1201 15:49:50.491243 4739 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 15:49:50 crc kubenswrapper[4739]: E1201 15:49:50.491270 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs podName:5f474df5-ced8-45ff-840a-c52898696bf7 nodeName:}" failed. No retries permitted until 2025-12-01 15:49:50.99126278 +0000 UTC m=+892.817008874 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs") pod "openstack-operator-controller-manager-59fc5775d4-chxkc" (UID: "5f474df5-ced8-45ff-840a-c52898696bf7") : secret "metrics-server-cert" not found Dec 01 15:49:50 crc kubenswrapper[4739]: W1201 15:49:50.520111 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44a9534e_febc_4a41_9ee0_af96baa95a78.slice/crio-2fb40facbc49e73f2084c800cbbe21cab6ad6a90bc727372c97f4a63edfd53cf WatchSource:0}: Error finding container 2fb40facbc49e73f2084c800cbbe21cab6ad6a90bc727372c97f4a63edfd53cf: Status 404 returned error can't find the container with id 2fb40facbc49e73f2084c800cbbe21cab6ad6a90bc727372c97f4a63edfd53cf Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.527130 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m969w\" (UniqueName: \"kubernetes.io/projected/47bf6f4a-97c7-4f1f-a00d-0c5151058f2c-kube-api-access-m969w\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qm8fw\" (UID: \"47bf6f4a-97c7-4f1f-a00d-0c5151058f2c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qm8fw" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.530940 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcmbv\" (UniqueName: \"kubernetes.io/projected/5f474df5-ced8-45ff-840a-c52898696bf7-kube-api-access-pcmbv\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.549004 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmntr"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.585757 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qm8fw" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.591386 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh\" (UID: \"81c50b9d-cef2-40b0-bb96-0e44676eaf93\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" Dec 01 15:49:50 crc kubenswrapper[4739]: E1201 15:49:50.593018 4739 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 15:49:50 crc kubenswrapper[4739]: E1201 15:49:50.593069 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert podName:81c50b9d-cef2-40b0-bb96-0e44676eaf93 nodeName:}" failed. No retries permitted until 2025-12-01 15:49:51.593050819 +0000 UTC m=+893.418796993 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" (UID: "81c50b9d-cef2-40b0-bb96-0e44676eaf93") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.658731 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.893025 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ll5cw"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.897296 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.907193 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ll5cw"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.976780 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-2m26g"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.981935 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-pqqzm"] Dec 01 15:49:50 crc kubenswrapper[4739]: I1201 15:49:50.991891 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-f2t8s"] Dec 01 15:49:50 crc kubenswrapper[4739]: W1201 15:49:50.995540 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8525c2bf_aedc_4816_8ea5_4ebdda160f3d.slice/crio-03b58b07f883fac95a18a74c58e058a3ec7ee1a456bb33e6172b6c327e7b3443 WatchSource:0}: Error finding container 03b58b07f883fac95a18a74c58e058a3ec7ee1a456bb33e6172b6c327e7b3443: Status 404 returned error can't find the container with id 03b58b07f883fac95a18a74c58e058a3ec7ee1a456bb33e6172b6c327e7b3443 Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.000859 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.001005 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-catalog-content\") pod \"certified-operators-ll5cw\" (UID: \"41c3fd88-c7a2-4741-a3da-6b9f33b42b46\") " pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.001068 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.001102 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mzzj\" (UniqueName: \"kubernetes.io/projected/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-kube-api-access-6mzzj\") pod \"certified-operators-ll5cw\" (UID: \"41c3fd88-c7a2-4741-a3da-6b9f33b42b46\") " pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.001128 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-utilities\") pod \"certified-operators-ll5cw\" (UID: \"41c3fd88-c7a2-4741-a3da-6b9f33b42b46\") " pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.001336 4739 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.001386 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs podName:5f474df5-ced8-45ff-840a-c52898696bf7 nodeName:}" failed. No retries permitted until 2025-12-01 15:49:52.00136874 +0000 UTC m=+893.827114844 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs") pod "openstack-operator-controller-manager-59fc5775d4-chxkc" (UID: "5f474df5-ced8-45ff-840a-c52898696bf7") : secret "metrics-server-cert" not found Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.001517 4739 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.001545 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs podName:5f474df5-ced8-45ff-840a-c52898696bf7 nodeName:}" failed. No retries permitted until 2025-12-01 15:49:52.001536826 +0000 UTC m=+893.827282920 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs") pod "openstack-operator-controller-manager-59fc5775d4-chxkc" (UID: "5f474df5-ced8-45ff-840a-c52898696bf7") : secret "webhook-server-cert" not found Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.002234 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-d2dxq"] Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.026881 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd"] Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.028972 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmntr" event={"ID":"3422b21c-7fef-4c3a-8990-85fa0d40b027","Type":"ContainerStarted","Data":"fee2e145c05b816a635f11c247474829d960a7e0b9c701ff31bcc3f21efb67eb"} Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.030335 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nz6dz" event={"ID":"97f978b1-c0ba-4fa7-8a8f-91de3ab80bc5","Type":"ContainerStarted","Data":"db70ce722a1fb64e0bf1f986131900daff7b08c4cd30c38ea6200bd120b2b854"} Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.034280 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g7bc7"] Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.034850 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-pqqzm" event={"ID":"ebe9e6cb-7cc7-4a43-ba59-ff0b8209d314","Type":"ContainerStarted","Data":"fee8e47beabb433e33e7d1a5f0d47235b03c9d5a58e79541db3ec42f6788c92a"} Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.044123 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-th7k2" event={"ID":"44a9534e-febc-4a41-9ee0-af96baa95a78","Type":"ContainerStarted","Data":"2fb40facbc49e73f2084c800cbbe21cab6ad6a90bc727372c97f4a63edfd53cf"} Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.045853 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-2m26g" event={"ID":"8525c2bf-aedc-4816-8ea5-4ebdda160f3d","Type":"ContainerStarted","Data":"03b58b07f883fac95a18a74c58e058a3ec7ee1a456bb33e6172b6c327e7b3443"} Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.052301 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq" event={"ID":"28db1029-7abe-4d58-afbd-0022422c286c","Type":"ContainerStarted","Data":"f28e6e3f7a636d6ea3a4b0e0ae0196589d0de0869249c9e771fc7f70a59a84c9"} Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.100973 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt"] Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.102087 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mzzj\" (UniqueName: \"kubernetes.io/projected/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-kube-api-access-6mzzj\") pod \"certified-operators-ll5cw\" (UID: \"41c3fd88-c7a2-4741-a3da-6b9f33b42b46\") " pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.102117 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-utilities\") pod \"certified-operators-ll5cw\" (UID: \"41c3fd88-c7a2-4741-a3da-6b9f33b42b46\") " pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.102207 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-catalog-content\") pod \"certified-operators-ll5cw\" (UID: \"41c3fd88-c7a2-4741-a3da-6b9f33b42b46\") " pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.102931 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-catalog-content\") pod \"certified-operators-ll5cw\" (UID: \"41c3fd88-c7a2-4741-a3da-6b9f33b42b46\") " pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.103014 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-utilities\") pod \"certified-operators-ll5cw\" (UID: \"41c3fd88-c7a2-4741-a3da-6b9f33b42b46\") " pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:49:51 crc kubenswrapper[4739]: W1201 15:49:51.119780 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod974d1918_6e6a_4660_898c_b8d8db439f46.slice/crio-2d7a8a8aeab8b0a2b2b8f1c18086e9e6b966b4979025683af5176dd08f65304a WatchSource:0}: Error finding container 2d7a8a8aeab8b0a2b2b8f1c18086e9e6b966b4979025683af5176dd08f65304a: Status 404 returned error can't find the container with id 2d7a8a8aeab8b0a2b2b8f1c18086e9e6b966b4979025683af5176dd08f65304a Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.122369 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-4nwgp"] Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.129572 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mzzj\" (UniqueName: \"kubernetes.io/projected/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-kube-api-access-6mzzj\") pod \"certified-operators-ll5cw\" (UID: \"41c3fd88-c7a2-4741-a3da-6b9f33b42b46\") " pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:49:51 crc kubenswrapper[4739]: W1201 15:49:51.135472 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d6c8b0a_e856_4794_a3ce_f72e1f6d0c91.slice/crio-8c6ffea9c3c41b4fbc41d9b75bc0c0e0561d1780975a10fa97660c5c76a95cb2 WatchSource:0}: Error finding container 8c6ffea9c3c41b4fbc41d9b75bc0c0e0561d1780975a10fa97660c5c76a95cb2: Status 404 returned error can't find the container with id 8c6ffea9c3c41b4fbc41d9b75bc0c0e0561d1780975a10fa97660c5c76a95cb2 Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.137950 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jrc6n"] Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.214491 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.239258 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-zrhx2"] Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.259042 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt"] Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.305097 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-kkchm"] Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.305996 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert\") pod \"infra-operator-controller-manager-57548d458d-gpkcn\" (UID: \"54dcd8b3-40f6-4487-bc4f-810c6aebb6fe\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.306253 4739 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.306300 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert podName:54dcd8b3-40f6-4487-bc4f-810c6aebb6fe nodeName:}" failed. No retries permitted until 2025-12-01 15:49:53.306284454 +0000 UTC m=+895.132030548 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert") pod "infra-operator-controller-manager-57548d458d-gpkcn" (UID: "54dcd8b3-40f6-4487-bc4f-810c6aebb6fe") : secret "infra-operator-webhook-server-cert" not found Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.309272 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm"] Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.333907 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-54nx7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-fxvzt_openstack-operators(f60761e0-1afe-46fa-90fc-5ed9f42f730d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.344285 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-54nx7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-fxvzt_openstack-operators(f60761e0-1afe-46fa-90fc-5ed9f42f730d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.345369 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt" podUID="f60761e0-1afe-46fa-90fc-5ed9f42f730d" Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.347065 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9227w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-kkchm_openstack-operators(a4706184-6b4f-4416-b9d6-430087d5577c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.347168 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lm8qg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-4pwqm_openstack-operators(bed989ab-cd94-41f1-91ea-8c1bdfc056af): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.355843 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9227w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-kkchm_openstack-operators(a4706184-6b4f-4416-b9d6-430087d5577c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.356245 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lm8qg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-4pwqm_openstack-operators(bed989ab-cd94-41f1-91ea-8c1bdfc056af): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.359363 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm" podUID="bed989ab-cd94-41f1-91ea-8c1bdfc056af" Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.361506 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkchm" podUID="a4706184-6b4f-4416-b9d6-430087d5577c" Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.417603 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-px5zx"] Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.491071 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m969w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-qm8fw_openstack-operators(47bf6f4a-97c7-4f1f-a00d-0c5151058f2c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.492206 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qm8fw" podUID="47bf6f4a-97c7-4f1f-a00d-0c5151058f2c" Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.492270 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qm8fw"] Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.502891 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt"] Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.520820 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9g49c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-xgpkt_openstack-operators(7ff65da3-52d5-4e80-9f48-d23d3bc22ff6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.523785 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9g49c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-xgpkt_openstack-operators(7ff65da3-52d5-4e80-9f48-d23d3bc22ff6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.526763 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt" podUID="7ff65da3-52d5-4e80-9f48-d23d3bc22ff6" Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.615074 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh\" (UID: \"81c50b9d-cef2-40b0-bb96-0e44676eaf93\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.615226 4739 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 15:49:51 crc kubenswrapper[4739]: E1201 15:49:51.615287 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert podName:81c50b9d-cef2-40b0-bb96-0e44676eaf93 nodeName:}" failed. No retries permitted until 2025-12-01 15:49:53.615270201 +0000 UTC m=+895.441016295 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" (UID: "81c50b9d-cef2-40b0-bb96-0e44676eaf93") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 15:49:51 crc kubenswrapper[4739]: I1201 15:49:51.690906 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ll5cw"] Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.020435 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.020549 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:52 crc kubenswrapper[4739]: E1201 15:49:52.020641 4739 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 15:49:52 crc kubenswrapper[4739]: E1201 15:49:52.020726 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs podName:5f474df5-ced8-45ff-840a-c52898696bf7 nodeName:}" failed. No retries permitted until 2025-12-01 15:49:54.020704926 +0000 UTC m=+895.846451100 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs") pod "openstack-operator-controller-manager-59fc5775d4-chxkc" (UID: "5f474df5-ced8-45ff-840a-c52898696bf7") : secret "metrics-server-cert" not found Dec 01 15:49:52 crc kubenswrapper[4739]: E1201 15:49:52.020743 4739 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 15:49:52 crc kubenswrapper[4739]: E1201 15:49:52.020802 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs podName:5f474df5-ced8-45ff-840a-c52898696bf7 nodeName:}" failed. No retries permitted until 2025-12-01 15:49:54.020783779 +0000 UTC m=+895.846529953 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs") pod "openstack-operator-controller-manager-59fc5775d4-chxkc" (UID: "5f474df5-ced8-45ff-840a-c52898696bf7") : secret "webhook-server-cert" not found Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.060809 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm" event={"ID":"bed989ab-cd94-41f1-91ea-8c1bdfc056af","Type":"ContainerStarted","Data":"19d58d8add340d5ea3bf6be5bf351aca20bb0c9331b511d714929d5813370a73"} Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.063331 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qm8fw" event={"ID":"47bf6f4a-97c7-4f1f-a00d-0c5151058f2c","Type":"ContainerStarted","Data":"f753b66fce2a11244592d47d84ec6092e12398024440140c8b5838e3f2b7798d"} Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.064902 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-f2t8s" event={"ID":"895e540c-7e60-45b1-8275-36588bb8e788","Type":"ContainerStarted","Data":"8944198dcee62d6cc9ee16d40731909a2e97dab8320c5b90d0e6f313564f9c09"} Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.066340 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-d2dxq" event={"ID":"7a27f279-7c57-439c-88f2-67914f2e7b9e","Type":"ContainerStarted","Data":"6ffdcc036ae3dcbe792b656d7ba73bfb64aa22c10a550086033cfbede6c2a960"} Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.069597 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jrc6n" event={"ID":"5d6c8b0a-e856-4794-a3ce-f72e1f6d0c91","Type":"ContainerStarted","Data":"8c6ffea9c3c41b4fbc41d9b75bc0c0e0561d1780975a10fa97660c5c76a95cb2"} Dec 01 15:49:52 crc kubenswrapper[4739]: E1201 15:49:52.070653 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qm8fw" podUID="47bf6f4a-97c7-4f1f-a00d-0c5151058f2c" Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.071561 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkchm" event={"ID":"a4706184-6b4f-4416-b9d6-430087d5577c","Type":"ContainerStarted","Data":"95451975392b842ed9e597f3c26bf124c945a3a3c05061fdaeafe3e7d9de20ad"} Dec 01 15:49:52 crc kubenswrapper[4739]: E1201 15:49:52.072259 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm" podUID="bed989ab-cd94-41f1-91ea-8c1bdfc056af" Dec 01 15:49:52 crc kubenswrapper[4739]: E1201 15:49:52.073067 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkchm" podUID="a4706184-6b4f-4416-b9d6-430087d5577c" Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.079864 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt" event={"ID":"7ff65da3-52d5-4e80-9f48-d23d3bc22ff6","Type":"ContainerStarted","Data":"d16292528ce3a40743885d50753554e6c4f0c09354bc29373b14283b115448fe"} Dec 01 15:49:52 crc kubenswrapper[4739]: E1201 15:49:52.081737 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt" podUID="7ff65da3-52d5-4e80-9f48-d23d3bc22ff6" Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.083048 4739 generic.go:334] "Generic (PLEG): container finished" podID="41c3fd88-c7a2-4741-a3da-6b9f33b42b46" containerID="bbae966eb5cf115e28348fd122e3a0921f0cd1e5dc9ae0811af2579fe7c3104a" exitCode=0 Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.083094 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ll5cw" event={"ID":"41c3fd88-c7a2-4741-a3da-6b9f33b42b46","Type":"ContainerDied","Data":"bbae966eb5cf115e28348fd122e3a0921f0cd1e5dc9ae0811af2579fe7c3104a"} Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.083133 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ll5cw" event={"ID":"41c3fd88-c7a2-4741-a3da-6b9f33b42b46","Type":"ContainerStarted","Data":"bf439f4ca6554eec30c5e8667a4ce3724a283bb03fa2a9a700644f966164066f"} Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.109257 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g7bc7" event={"ID":"644da74f-c26b-4fae-a81f-4091a1e95b87","Type":"ContainerStarted","Data":"52d8a5c19185f6b0cf43eb0264a21034b85a676e59a949d21e4e5ff1b71759ea"} Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.112137 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zrhx2" event={"ID":"c6ddb146-59dc-41c2-9278-466eda24296c","Type":"ContainerStarted","Data":"8b67af5a8208d01b58b564566225789c5e9b13d866a2a7014cb6e329e04f51a9"} Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.119320 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt" event={"ID":"974d1918-6e6a-4660-898c-b8d8db439f46","Type":"ContainerStarted","Data":"2d7a8a8aeab8b0a2b2b8f1c18086e9e6b966b4979025683af5176dd08f65304a"} Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.123936 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt" event={"ID":"f60761e0-1afe-46fa-90fc-5ed9f42f730d","Type":"ContainerStarted","Data":"b45fa2cc21eb3a36ef658e89546a175cc09b277e5fa5e132c299ea69a052f91a"} Dec 01 15:49:52 crc kubenswrapper[4739]: E1201 15:49:52.129407 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt" podUID="f60761e0-1afe-46fa-90fc-5ed9f42f730d" Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.130163 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4nwgp" event={"ID":"903f2159-9ae0-4b78-8447-3b5bd4a420ef","Type":"ContainerStarted","Data":"109c6c660d933437f031b8c256092bc6ad50c2a013859ac2b529947bda983482"} Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.139563 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-px5zx" event={"ID":"84c3dbfb-b9ac-4b9d-b942-e454b50bf132","Type":"ContainerStarted","Data":"d4838e9d7c728072fdb365883e9a15e3a6c9603d1adde3f3308249726e9ac55a"} Dec 01 15:49:52 crc kubenswrapper[4739]: I1201 15:49:52.140994 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd" event={"ID":"8f8c53db-885f-4548-bcc0-02ce36cd6543","Type":"ContainerStarted","Data":"294a6f3bb06b15858066e772c9f29748b0cae1de2a2b231fdd927760045b3f1d"} Dec 01 15:49:53 crc kubenswrapper[4739]: E1201 15:49:53.149731 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt" podUID="f60761e0-1afe-46fa-90fc-5ed9f42f730d" Dec 01 15:49:53 crc kubenswrapper[4739]: E1201 15:49:53.150056 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt" podUID="7ff65da3-52d5-4e80-9f48-d23d3bc22ff6" Dec 01 15:49:53 crc kubenswrapper[4739]: E1201 15:49:53.150102 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkchm" podUID="a4706184-6b4f-4416-b9d6-430087d5577c" Dec 01 15:49:53 crc kubenswrapper[4739]: E1201 15:49:53.150096 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qm8fw" podUID="47bf6f4a-97c7-4f1f-a00d-0c5151058f2c" Dec 01 15:49:53 crc kubenswrapper[4739]: E1201 15:49:53.150272 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm" podUID="bed989ab-cd94-41f1-91ea-8c1bdfc056af" Dec 01 15:49:53 crc kubenswrapper[4739]: I1201 15:49:53.337925 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert\") pod \"infra-operator-controller-manager-57548d458d-gpkcn\" (UID: \"54dcd8b3-40f6-4487-bc4f-810c6aebb6fe\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:49:53 crc kubenswrapper[4739]: E1201 15:49:53.338059 4739 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 15:49:53 crc kubenswrapper[4739]: E1201 15:49:53.338120 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert podName:54dcd8b3-40f6-4487-bc4f-810c6aebb6fe nodeName:}" failed. No retries permitted until 2025-12-01 15:49:57.3381029 +0000 UTC m=+899.163848994 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert") pod "infra-operator-controller-manager-57548d458d-gpkcn" (UID: "54dcd8b3-40f6-4487-bc4f-810c6aebb6fe") : secret "infra-operator-webhook-server-cert" not found Dec 01 15:49:53 crc kubenswrapper[4739]: I1201 15:49:53.642059 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh\" (UID: \"81c50b9d-cef2-40b0-bb96-0e44676eaf93\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" Dec 01 15:49:53 crc kubenswrapper[4739]: E1201 15:49:53.642245 4739 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 15:49:53 crc kubenswrapper[4739]: E1201 15:49:53.642323 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert podName:81c50b9d-cef2-40b0-bb96-0e44676eaf93 nodeName:}" failed. No retries permitted until 2025-12-01 15:49:57.642303813 +0000 UTC m=+899.468049907 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" (UID: "81c50b9d-cef2-40b0-bb96-0e44676eaf93") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 15:49:54 crc kubenswrapper[4739]: I1201 15:49:54.047300 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:54 crc kubenswrapper[4739]: I1201 15:49:54.047380 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:54 crc kubenswrapper[4739]: E1201 15:49:54.047436 4739 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 15:49:54 crc kubenswrapper[4739]: E1201 15:49:54.047504 4739 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 15:49:54 crc kubenswrapper[4739]: E1201 15:49:54.047556 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs podName:5f474df5-ced8-45ff-840a-c52898696bf7 nodeName:}" failed. No retries permitted until 2025-12-01 15:49:58.047537941 +0000 UTC m=+899.873284035 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs") pod "openstack-operator-controller-manager-59fc5775d4-chxkc" (UID: "5f474df5-ced8-45ff-840a-c52898696bf7") : secret "metrics-server-cert" not found Dec 01 15:49:54 crc kubenswrapper[4739]: E1201 15:49:54.047573 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs podName:5f474df5-ced8-45ff-840a-c52898696bf7 nodeName:}" failed. No retries permitted until 2025-12-01 15:49:58.047567052 +0000 UTC m=+899.873313146 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs") pod "openstack-operator-controller-manager-59fc5775d4-chxkc" (UID: "5f474df5-ced8-45ff-840a-c52898696bf7") : secret "webhook-server-cert" not found Dec 01 15:49:57 crc kubenswrapper[4739]: I1201 15:49:57.399535 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert\") pod \"infra-operator-controller-manager-57548d458d-gpkcn\" (UID: \"54dcd8b3-40f6-4487-bc4f-810c6aebb6fe\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:49:57 crc kubenswrapper[4739]: E1201 15:49:57.399800 4739 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 15:49:57 crc kubenswrapper[4739]: E1201 15:49:57.400093 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert podName:54dcd8b3-40f6-4487-bc4f-810c6aebb6fe nodeName:}" failed. No retries permitted until 2025-12-01 15:50:05.4000585 +0000 UTC m=+907.225804684 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert") pod "infra-operator-controller-manager-57548d458d-gpkcn" (UID: "54dcd8b3-40f6-4487-bc4f-810c6aebb6fe") : secret "infra-operator-webhook-server-cert" not found Dec 01 15:49:57 crc kubenswrapper[4739]: I1201 15:49:57.705168 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh\" (UID: \"81c50b9d-cef2-40b0-bb96-0e44676eaf93\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" Dec 01 15:49:57 crc kubenswrapper[4739]: E1201 15:49:57.705468 4739 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 15:49:57 crc kubenswrapper[4739]: E1201 15:49:57.705583 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert podName:81c50b9d-cef2-40b0-bb96-0e44676eaf93 nodeName:}" failed. No retries permitted until 2025-12-01 15:50:05.705552172 +0000 UTC m=+907.531298306 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" (UID: "81c50b9d-cef2-40b0-bb96-0e44676eaf93") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 15:49:58 crc kubenswrapper[4739]: I1201 15:49:58.111692 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:58 crc kubenswrapper[4739]: E1201 15:49:58.111976 4739 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 15:49:58 crc kubenswrapper[4739]: I1201 15:49:58.112445 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:49:58 crc kubenswrapper[4739]: E1201 15:49:58.112542 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs podName:5f474df5-ced8-45ff-840a-c52898696bf7 nodeName:}" failed. No retries permitted until 2025-12-01 15:50:06.112511142 +0000 UTC m=+907.938257266 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs") pod "openstack-operator-controller-manager-59fc5775d4-chxkc" (UID: "5f474df5-ced8-45ff-840a-c52898696bf7") : secret "webhook-server-cert" not found Dec 01 15:49:58 crc kubenswrapper[4739]: E1201 15:49:58.112653 4739 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 15:49:58 crc kubenswrapper[4739]: E1201 15:49:58.112746 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs podName:5f474df5-ced8-45ff-840a-c52898696bf7 nodeName:}" failed. No retries permitted until 2025-12-01 15:50:06.112721548 +0000 UTC m=+907.938467682 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs") pod "openstack-operator-controller-manager-59fc5775d4-chxkc" (UID: "5f474df5-ced8-45ff-840a-c52898696bf7") : secret "metrics-server-cert" not found Dec 01 15:50:05 crc kubenswrapper[4739]: I1201 15:50:05.441515 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert\") pod \"infra-operator-controller-manager-57548d458d-gpkcn\" (UID: \"54dcd8b3-40f6-4487-bc4f-810c6aebb6fe\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:50:05 crc kubenswrapper[4739]: E1201 15:50:05.441760 4739 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 15:50:05 crc kubenswrapper[4739]: E1201 15:50:05.442336 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert podName:54dcd8b3-40f6-4487-bc4f-810c6aebb6fe nodeName:}" failed. No retries permitted until 2025-12-01 15:50:21.442308076 +0000 UTC m=+923.268054210 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert") pod "infra-operator-controller-manager-57548d458d-gpkcn" (UID: "54dcd8b3-40f6-4487-bc4f-810c6aebb6fe") : secret "infra-operator-webhook-server-cert" not found Dec 01 15:50:05 crc kubenswrapper[4739]: I1201 15:50:05.747679 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh\" (UID: \"81c50b9d-cef2-40b0-bb96-0e44676eaf93\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" Dec 01 15:50:05 crc kubenswrapper[4739]: I1201 15:50:05.765189 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/81c50b9d-cef2-40b0-bb96-0e44676eaf93-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh\" (UID: \"81c50b9d-cef2-40b0-bb96-0e44676eaf93\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" Dec 01 15:50:05 crc kubenswrapper[4739]: I1201 15:50:05.899701 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-9fspg" Dec 01 15:50:05 crc kubenswrapper[4739]: I1201 15:50:05.908134 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" Dec 01 15:50:06 crc kubenswrapper[4739]: I1201 15:50:06.153073 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:50:06 crc kubenswrapper[4739]: I1201 15:50:06.153258 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:50:06 crc kubenswrapper[4739]: E1201 15:50:06.153328 4739 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 15:50:06 crc kubenswrapper[4739]: E1201 15:50:06.153466 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs podName:5f474df5-ced8-45ff-840a-c52898696bf7 nodeName:}" failed. No retries permitted until 2025-12-01 15:50:22.153438888 +0000 UTC m=+923.979185012 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs") pod "openstack-operator-controller-manager-59fc5775d4-chxkc" (UID: "5f474df5-ced8-45ff-840a-c52898696bf7") : secret "webhook-server-cert" not found Dec 01 15:50:06 crc kubenswrapper[4739]: I1201 15:50:06.166479 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-metrics-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:50:09 crc kubenswrapper[4739]: I1201 15:50:09.622806 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:50:09 crc kubenswrapper[4739]: I1201 15:50:09.623347 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:50:10 crc kubenswrapper[4739]: E1201 15:50:10.986694 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 01 15:50:10 crc kubenswrapper[4739]: E1201 15:50:10.988208 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b26mh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-4nwgp_openstack-operators(903f2159-9ae0-4b78-8447-3b5bd4a420ef): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 15:50:11 crc kubenswrapper[4739]: E1201 15:50:11.643241 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:440cde33d3a2a0c545cd1c110a3634eb85544370f448865b97a13c38034b0172" Dec 01 15:50:11 crc kubenswrapper[4739]: E1201 15:50:11.643710 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:440cde33d3a2a0c545cd1c110a3634eb85544370f448865b97a13c38034b0172,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ccht4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-668d9c48b9-2m26g_openstack-operators(8525c2bf-aedc-4816-8ea5-4ebdda160f3d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 15:50:12 crc kubenswrapper[4739]: E1201 15:50:12.189051 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 01 15:50:12 crc kubenswrapper[4739]: E1201 15:50:12.189218 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jsz6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-zrhx2_openstack-operators(c6ddb146-59dc-41c2-9278-466eda24296c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 15:50:12 crc kubenswrapper[4739]: E1201 15:50:12.830366 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 01 15:50:12 crc kubenswrapper[4739]: E1201 15:50:12.830844 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xk8jk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-8bsmt_openstack-operators(974d1918-6e6a-4660-898c-b8d8db439f46): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 15:50:12 crc kubenswrapper[4739]: E1201 15:50:12.875727 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.44:5001/openstack-k8s-operators/cinder-operator:509169b1ebe024d91d4a41c1268d248d1a70afec" Dec 01 15:50:12 crc kubenswrapper[4739]: E1201 15:50:12.875807 4739 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.44:5001/openstack-k8s-operators/cinder-operator:509169b1ebe024d91d4a41c1268d248d1a70afec" Dec 01 15:50:12 crc kubenswrapper[4739]: E1201 15:50:12.875978 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.44:5001/openstack-k8s-operators/cinder-operator:509169b1ebe024d91d4a41c1268d248d1a70afec,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tdht5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-5d986d85c9-2s7sq_openstack-operators(28db1029-7abe-4d58-afbd-0022422c286c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 15:50:13 crc kubenswrapper[4739]: E1201 15:50:13.338356 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3" Dec 01 15:50:13 crc kubenswrapper[4739]: E1201 15:50:13.338775 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6cdnw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-546d4bdf48-758wd_openstack-operators(8f8c53db-885f-4548-bcc0-02ce36cd6543): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 15:50:17 crc kubenswrapper[4739]: I1201 15:50:17.178586 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh"] Dec 01 15:50:17 crc kubenswrapper[4739]: I1201 15:50:17.317882 4739 generic.go:334] "Generic (PLEG): container finished" podID="41c3fd88-c7a2-4741-a3da-6b9f33b42b46" containerID="fe05f1a56b1e6cf55406a435c84d8609e7fa5e5e6855a05aeda63fdc6a68f1c9" exitCode=0 Dec 01 15:50:17 crc kubenswrapper[4739]: I1201 15:50:17.317974 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ll5cw" event={"ID":"41c3fd88-c7a2-4741-a3da-6b9f33b42b46","Type":"ContainerDied","Data":"fe05f1a56b1e6cf55406a435c84d8609e7fa5e5e6855a05aeda63fdc6a68f1c9"} Dec 01 15:50:17 crc kubenswrapper[4739]: I1201 15:50:17.320096 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmntr" event={"ID":"3422b21c-7fef-4c3a-8990-85fa0d40b027","Type":"ContainerStarted","Data":"36a1441efff4059f30b47a107d459d587b4ec00593666d0ec1a1d7c7ee95675b"} Dec 01 15:50:17 crc kubenswrapper[4739]: W1201 15:50:17.518271 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81c50b9d_cef2_40b0_bb96_0e44676eaf93.slice/crio-cbfda0204c9e61574739657af496fa2509d56eeb1b2073a232b7b5c7ae288416 WatchSource:0}: Error finding container cbfda0204c9e61574739657af496fa2509d56eeb1b2073a232b7b5c7ae288416: Status 404 returned error can't find the container with id cbfda0204c9e61574739657af496fa2509d56eeb1b2073a232b7b5c7ae288416 Dec 01 15:50:18 crc kubenswrapper[4739]: I1201 15:50:18.334576 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g7bc7" event={"ID":"644da74f-c26b-4fae-a81f-4091a1e95b87","Type":"ContainerStarted","Data":"6344b2f4370dc762508f2b2127415f325b56c4292d497bb46d6fb3fe9e7f9506"} Dec 01 15:50:18 crc kubenswrapper[4739]: I1201 15:50:18.338034 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm" event={"ID":"bed989ab-cd94-41f1-91ea-8c1bdfc056af","Type":"ContainerStarted","Data":"b9813f3173b46717c74525d943e6c7d25e340e9eef772bf7f8e88d89cad1c4bf"} Dec 01 15:50:18 crc kubenswrapper[4739]: I1201 15:50:18.366495 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-f2t8s" event={"ID":"895e540c-7e60-45b1-8275-36588bb8e788","Type":"ContainerStarted","Data":"f11c060568b03ee5189ec380920024b39e2ed4b4ecc9ae06f131a5510ef29ac8"} Dec 01 15:50:18 crc kubenswrapper[4739]: I1201 15:50:18.376033 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt" event={"ID":"7ff65da3-52d5-4e80-9f48-d23d3bc22ff6","Type":"ContainerStarted","Data":"8b69af13e4da06593a4261a72552597f762c2f6cb69fac92cee021f455442f36"} Dec 01 15:50:18 crc kubenswrapper[4739]: I1201 15:50:18.377560 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jrc6n" event={"ID":"5d6c8b0a-e856-4794-a3ce-f72e1f6d0c91","Type":"ContainerStarted","Data":"6b3eb29c1646cc1d880f7a51b78bbbc5c6da979e9c50e866f588e3098e0f08f5"} Dec 01 15:50:18 crc kubenswrapper[4739]: I1201 15:50:18.378346 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkchm" event={"ID":"a4706184-6b4f-4416-b9d6-430087d5577c","Type":"ContainerStarted","Data":"4aad92b34357eeb0d21f25e89bdcb3264ae6f3380c1822323e7f3d3c33531db6"} Dec 01 15:50:18 crc kubenswrapper[4739]: I1201 15:50:18.378968 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" event={"ID":"81c50b9d-cef2-40b0-bb96-0e44676eaf93","Type":"ContainerStarted","Data":"cbfda0204c9e61574739657af496fa2509d56eeb1b2073a232b7b5c7ae288416"} Dec 01 15:50:18 crc kubenswrapper[4739]: I1201 15:50:18.385841 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-th7k2" event={"ID":"44a9534e-febc-4a41-9ee0-af96baa95a78","Type":"ContainerStarted","Data":"1c1c886988fd9b825f2b090ebec2eb47de1455a5db21b90d08b5578cd551436e"} Dec 01 15:50:18 crc kubenswrapper[4739]: I1201 15:50:18.394298 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-d2dxq" event={"ID":"7a27f279-7c57-439c-88f2-67914f2e7b9e","Type":"ContainerStarted","Data":"4268451bac44d7585ca987109ce57b8fb6dab79bf3745e96c6185c65227818d6"} Dec 01 15:50:18 crc kubenswrapper[4739]: I1201 15:50:18.430703 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nz6dz" event={"ID":"97f978b1-c0ba-4fa7-8a8f-91de3ab80bc5","Type":"ContainerStarted","Data":"e9a0f7d5b89931fb695dbeff8cd14f7c77ab278a8488011eb3a380dd8654a7fb"} Dec 01 15:50:18 crc kubenswrapper[4739]: I1201 15:50:18.447444 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-pqqzm" event={"ID":"ebe9e6cb-7cc7-4a43-ba59-ff0b8209d314","Type":"ContainerStarted","Data":"72ba786d701c52ea01b1756aea408469b88552daa2b37761e7899330aa29983f"} Dec 01 15:50:18 crc kubenswrapper[4739]: I1201 15:50:18.466942 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-px5zx" event={"ID":"84c3dbfb-b9ac-4b9d-b942-e454b50bf132","Type":"ContainerStarted","Data":"c030598a964cb3084284727a0b0267c8b537e99827566064a13b416afb31ef9f"} Dec 01 15:50:19 crc kubenswrapper[4739]: I1201 15:50:19.492958 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt" event={"ID":"f60761e0-1afe-46fa-90fc-5ed9f42f730d","Type":"ContainerStarted","Data":"0fd2d1b8e63d68e5b94540b33486a18ecc9916bafa39af3f4a2422d5d2cc6e82"} Dec 01 15:50:20 crc kubenswrapper[4739]: I1201 15:50:20.502647 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qm8fw" event={"ID":"47bf6f4a-97c7-4f1f-a00d-0c5151058f2c","Type":"ContainerStarted","Data":"ed51e4b4fc3550ecbdf77f1d99ec80816e5d2cb19a53f783a5a8ffe9ea9bd70e"} Dec 01 15:50:20 crc kubenswrapper[4739]: I1201 15:50:20.525838 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qm8fw" podStartSLOduration=4.96494297 podStartE2EDuration="30.525817005s" podCreationTimestamp="2025-12-01 15:49:50 +0000 UTC" firstStartedPulling="2025-12-01 15:49:51.490924184 +0000 UTC m=+893.316670278" lastFinishedPulling="2025-12-01 15:50:17.051798209 +0000 UTC m=+918.877544313" observedRunningTime="2025-12-01 15:50:20.516380036 +0000 UTC m=+922.342126130" watchObservedRunningTime="2025-12-01 15:50:20.525817005 +0000 UTC m=+922.351563109" Dec 01 15:50:21 crc kubenswrapper[4739]: E1201 15:50:21.375629 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt" podUID="974d1918-6e6a-4660-898c-b8d8db439f46" Dec 01 15:50:21 crc kubenswrapper[4739]: E1201 15:50:21.486038 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd" podUID="8f8c53db-885f-4548-bcc0-02ce36cd6543" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.493854 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert\") pod \"infra-operator-controller-manager-57548d458d-gpkcn\" (UID: \"54dcd8b3-40f6-4487-bc4f-810c6aebb6fe\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.505452 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/54dcd8b3-40f6-4487-bc4f-810c6aebb6fe-cert\") pod \"infra-operator-controller-manager-57548d458d-gpkcn\" (UID: \"54dcd8b3-40f6-4487-bc4f-810c6aebb6fe\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.538639 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-f2t8s" event={"ID":"895e540c-7e60-45b1-8275-36588bb8e788","Type":"ContainerStarted","Data":"33aec067a2cd313dd303ba8b07ebfdb7f797c2a62932774079711d55eb694ded"} Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.539135 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-f2t8s" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.543412 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmntr" event={"ID":"3422b21c-7fef-4c3a-8990-85fa0d40b027","Type":"ContainerStarted","Data":"148ca96a35cb9793f53bc0fba529d7456a4b19b16becf80a039047fd13563e5d"} Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.544068 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmntr" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.544929 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" event={"ID":"81c50b9d-cef2-40b0-bb96-0e44676eaf93","Type":"ContainerStarted","Data":"c44c7a8c9681413bbd8f56dbb3e4efba015be07fc694d280931ce2f79a544be9"} Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.552808 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ll5cw" event={"ID":"41c3fd88-c7a2-4741-a3da-6b9f33b42b46","Type":"ContainerStarted","Data":"35cc6cc2101ab475c7d8b59b8b43719f372c6d7020913f2e849277da0f049a3e"} Dec 01 15:50:21 crc kubenswrapper[4739]: E1201 15:50:21.563694 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4nwgp" podUID="903f2159-9ae0-4b78-8447-3b5bd4a420ef" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.571046 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-f2t8s" podStartSLOduration=2.866594663 podStartE2EDuration="32.571029106s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:51.044650672 +0000 UTC m=+892.870396766" lastFinishedPulling="2025-12-01 15:50:20.749085105 +0000 UTC m=+922.574831209" observedRunningTime="2025-12-01 15:50:21.570636644 +0000 UTC m=+923.396382738" watchObservedRunningTime="2025-12-01 15:50:21.571029106 +0000 UTC m=+923.396775200" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.572367 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt" event={"ID":"974d1918-6e6a-4660-898c-b8d8db439f46","Type":"ContainerStarted","Data":"1caa7fc575be0eaf597ffb95d71e5d786f02361dd3c73258cd163d445d5ca605"} Dec 01 15:50:21 crc kubenswrapper[4739]: E1201 15:50:21.575745 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt" podUID="974d1918-6e6a-4660-898c-b8d8db439f46" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.576818 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm" event={"ID":"bed989ab-cd94-41f1-91ea-8c1bdfc056af","Type":"ContainerStarted","Data":"e6f71e0af8f1b16a1352a2648bc450f8287e03ab1769283c60ab6ff28d59bd7a"} Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.577390 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.578360 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd" event={"ID":"8f8c53db-885f-4548-bcc0-02ce36cd6543","Type":"ContainerStarted","Data":"266b54f7942f05138af7f5191e3577af54e96f17e454ae60e23b5f755c3b1121"} Dec 01 15:50:21 crc kubenswrapper[4739]: E1201 15:50:21.579281 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd" podUID="8f8c53db-885f-4548-bcc0-02ce36cd6543" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.580256 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt" event={"ID":"f60761e0-1afe-46fa-90fc-5ed9f42f730d","Type":"ContainerStarted","Data":"edd40e40445e937974b5785d03b78b6a17f4799d5580e064bd0d13b6dc9d1194"} Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.580632 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.596641 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-th7k2" event={"ID":"44a9534e-febc-4a41-9ee0-af96baa95a78","Type":"ContainerStarted","Data":"6ac902f316fac6abed1b5b12ae92a7e32210896c96e66ddddf6b8607c99436f7"} Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.597253 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-th7k2" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.597962 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmntr" podStartSLOduration=2.495944196 podStartE2EDuration="32.597953373s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:50.609002248 +0000 UTC m=+892.434748342" lastFinishedPulling="2025-12-01 15:50:20.711011425 +0000 UTC m=+922.536757519" observedRunningTime="2025-12-01 15:50:21.596519089 +0000 UTC m=+923.422265183" watchObservedRunningTime="2025-12-01 15:50:21.597953373 +0000 UTC m=+923.423699467" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.630462 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ll5cw" podStartSLOduration=2.918536915 podStartE2EDuration="31.630442591s" podCreationTimestamp="2025-12-01 15:49:50 +0000 UTC" firstStartedPulling="2025-12-01 15:49:52.08473551 +0000 UTC m=+893.910481604" lastFinishedPulling="2025-12-01 15:50:20.796641186 +0000 UTC m=+922.622387280" observedRunningTime="2025-12-01 15:50:21.627182821 +0000 UTC m=+923.452928925" watchObservedRunningTime="2025-12-01 15:50:21.630442591 +0000 UTC m=+923.456188685" Dec 01 15:50:21 crc kubenswrapper[4739]: E1201 15:50:21.644710 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq" podUID="28db1029-7abe-4d58-afbd-0022422c286c" Dec 01 15:50:21 crc kubenswrapper[4739]: E1201 15:50:21.645210 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-2m26g" podUID="8525c2bf-aedc-4816-8ea5-4ebdda160f3d" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.678244 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm" podStartSLOduration=3.263330129 podStartE2EDuration="32.678224179s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:51.347114872 +0000 UTC m=+893.172860966" lastFinishedPulling="2025-12-01 15:50:20.762008922 +0000 UTC m=+922.587755016" observedRunningTime="2025-12-01 15:50:21.6753364 +0000 UTC m=+923.501082494" watchObservedRunningTime="2025-12-01 15:50:21.678224179 +0000 UTC m=+923.503970273" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.685054 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-th7k2" podStartSLOduration=2.499019876 podStartE2EDuration="32.685034978s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:50.525121027 +0000 UTC m=+892.350867121" lastFinishedPulling="2025-12-01 15:50:20.711136109 +0000 UTC m=+922.536882223" observedRunningTime="2025-12-01 15:50:21.656498381 +0000 UTC m=+923.482244475" watchObservedRunningTime="2025-12-01 15:50:21.685034978 +0000 UTC m=+923.510781072" Dec 01 15:50:21 crc kubenswrapper[4739]: E1201 15:50:21.744044 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zrhx2" podUID="c6ddb146-59dc-41c2-9278-466eda24296c" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.757391 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt" podStartSLOduration=3.191543833 podStartE2EDuration="32.757374s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:51.333753331 +0000 UTC m=+893.159499415" lastFinishedPulling="2025-12-01 15:50:20.899583498 +0000 UTC m=+922.725329582" observedRunningTime="2025-12-01 15:50:21.72315791 +0000 UTC m=+923.548904024" watchObservedRunningTime="2025-12-01 15:50:21.757374 +0000 UTC m=+923.583120094" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.761701 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-gwc9m" Dec 01 15:50:21 crc kubenswrapper[4739]: I1201 15:50:21.768657 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.211425 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.220240 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5f474df5-ced8-45ff-840a-c52898696bf7-webhook-certs\") pod \"openstack-operator-controller-manager-59fc5775d4-chxkc\" (UID: \"5f474df5-ced8-45ff-840a-c52898696bf7\") " pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.326182 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-qscx2" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.335294 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.386369 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn"] Dec 01 15:50:22 crc kubenswrapper[4739]: W1201 15:50:22.414837 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54dcd8b3_40f6_4487_bc4f_810c6aebb6fe.slice/crio-4024051a23656ec1f708b658cdb984dc770f0686ef2d3dccda6cb9ae776d65ba WatchSource:0}: Error finding container 4024051a23656ec1f708b658cdb984dc770f0686ef2d3dccda6cb9ae776d65ba: Status 404 returned error can't find the container with id 4024051a23656ec1f708b658cdb984dc770f0686ef2d3dccda6cb9ae776d65ba Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.604124 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" event={"ID":"54dcd8b3-40f6-4487-bc4f-810c6aebb6fe","Type":"ContainerStarted","Data":"4024051a23656ec1f708b658cdb984dc770f0686ef2d3dccda6cb9ae776d65ba"} Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.610702 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jrc6n" event={"ID":"5d6c8b0a-e856-4794-a3ce-f72e1f6d0c91","Type":"ContainerStarted","Data":"779e020a1b1b2cc35c8cc8fea3d63846f04d42c3248550695af88a7b42de8ad3"} Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.610907 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jrc6n" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.613335 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jrc6n" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.618045 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkchm" event={"ID":"a4706184-6b4f-4416-b9d6-430087d5577c","Type":"ContainerStarted","Data":"1ff78aed33512112e3ded16824c3972ecd558fc590a15993cdebffb4417d786f"} Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.618182 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkchm" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.619713 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" event={"ID":"81c50b9d-cef2-40b0-bb96-0e44676eaf93","Type":"ContainerStarted","Data":"5691e4726b80386f2ec294c313968e150e7d9d50eba98c55ef6a5d804e7f09ca"} Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.619832 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.624974 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkchm" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.626403 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g7bc7" event={"ID":"644da74f-c26b-4fae-a81f-4091a1e95b87","Type":"ContainerStarted","Data":"a09dbc1ad93ad827c798cf9b60e8b62963d178c28a2923396920dc333d64414e"} Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.626853 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g7bc7" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.634196 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-px5zx" event={"ID":"84c3dbfb-b9ac-4b9d-b942-e454b50bf132","Type":"ContainerStarted","Data":"a7631318dbf883e471db4d235ffb5124ccec30333f1f4854258efe22d2120ef8"} Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.634607 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-px5zx" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.634813 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-jrc6n" podStartSLOduration=3.849364492 podStartE2EDuration="33.634793606s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:51.138270725 +0000 UTC m=+892.964016809" lastFinishedPulling="2025-12-01 15:50:20.923699829 +0000 UTC m=+922.749445923" observedRunningTime="2025-12-01 15:50:22.631750673 +0000 UTC m=+924.457496767" watchObservedRunningTime="2025-12-01 15:50:22.634793606 +0000 UTC m=+924.460539700" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.635050 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g7bc7" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.636158 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zrhx2" event={"ID":"c6ddb146-59dc-41c2-9278-466eda24296c","Type":"ContainerStarted","Data":"76cae5ab9407adbd3bca352ca4267773f63fa4f6a6ce59c0c8f99a3ab36fb0fc"} Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.640016 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-d2dxq" event={"ID":"7a27f279-7c57-439c-88f2-67914f2e7b9e","Type":"ContainerStarted","Data":"83509024a6020a635ead576827bc282441608ea655653834e0e82802c306ebcf"} Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.641215 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-d2dxq" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.641973 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-px5zx" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.643030 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-d2dxq" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.645461 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nz6dz" event={"ID":"97f978b1-c0ba-4fa7-8a8f-91de3ab80bc5","Type":"ContainerStarted","Data":"20e15fb6f1a9efd42c3dc915275bfa5e421bb06c2960ac98c8b45e4699794d30"} Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.646218 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nz6dz" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.655629 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nz6dz" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.656579 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-pqqzm" event={"ID":"ebe9e6cb-7cc7-4a43-ba59-ff0b8209d314","Type":"ContainerStarted","Data":"e8f645c874bab56e2782186c5ab70cbfdd324edb45a52100ba37dde54ce08775"} Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.656604 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-pqqzm" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.659960 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-pqqzm" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.661236 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq" event={"ID":"28db1029-7abe-4d58-afbd-0022422c286c","Type":"ContainerStarted","Data":"6d94ddd72c7f0da1d08ac0ebee36678bb8fb9cab6e828ab03aa1f19e77757b0e"} Dec 01 15:50:22 crc kubenswrapper[4739]: E1201 15:50:22.663247 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.44:5001/openstack-k8s-operators/cinder-operator:509169b1ebe024d91d4a41c1268d248d1a70afec\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq" podUID="28db1029-7abe-4d58-afbd-0022422c286c" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.666367 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4nwgp" event={"ID":"903f2159-9ae0-4b78-8447-3b5bd4a420ef","Type":"ContainerStarted","Data":"a4ece02110f1fa4da53426b026ee9e9f68eb6723e2eb37eb78cc751e4d82ddf1"} Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.676798 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt" event={"ID":"7ff65da3-52d5-4e80-9f48-d23d3bc22ff6","Type":"ContainerStarted","Data":"4732373c358b45e811aaaa7a8e6e62b96520fc198b85347eb54ea3a497fbceaa"} Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.677775 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.680460 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-2m26g" event={"ID":"8525c2bf-aedc-4816-8ea5-4ebdda160f3d","Type":"ContainerStarted","Data":"7b920727124fd1b6fcef2e7f71ef295714423d066789512ff297c566b0a20f69"} Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.681726 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt" Dec 01 15:50:22 crc kubenswrapper[4739]: E1201 15:50:22.681951 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd" podUID="8f8c53db-885f-4548-bcc0-02ce36cd6543" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.682846 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-f2t8s" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.687166 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" podStartSLOduration=30.468023019 podStartE2EDuration="33.687150454s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:50:17.520801268 +0000 UTC m=+919.346547362" lastFinishedPulling="2025-12-01 15:50:20.739928693 +0000 UTC m=+922.565674797" observedRunningTime="2025-12-01 15:50:22.679481369 +0000 UTC m=+924.505227463" watchObservedRunningTime="2025-12-01 15:50:22.687150454 +0000 UTC m=+924.512896548" Dec 01 15:50:22 crc kubenswrapper[4739]: E1201 15:50:22.691131 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt" podUID="974d1918-6e6a-4660-898c-b8d8db439f46" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.697148 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-th7k2" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.707596 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkchm" podStartSLOduration=4.130630814 podStartE2EDuration="33.707455078s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:51.346955487 +0000 UTC m=+893.172701581" lastFinishedPulling="2025-12-01 15:50:20.923779751 +0000 UTC m=+922.749525845" observedRunningTime="2025-12-01 15:50:22.700838925 +0000 UTC m=+924.526585019" watchObservedRunningTime="2025-12-01 15:50:22.707455078 +0000 UTC m=+924.533201172" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.710593 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmntr" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.723198 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-g7bc7" podStartSLOduration=3.881023128 podStartE2EDuration="33.723179151s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:51.05890779 +0000 UTC m=+892.884653884" lastFinishedPulling="2025-12-01 15:50:20.901063813 +0000 UTC m=+922.726809907" observedRunningTime="2025-12-01 15:50:22.722031486 +0000 UTC m=+924.547777580" watchObservedRunningTime="2025-12-01 15:50:22.723179151 +0000 UTC m=+924.548925245" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.774847 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nz6dz" podStartSLOduration=3.400294973 podStartE2EDuration="33.774828028s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:50.422131482 +0000 UTC m=+892.247877576" lastFinishedPulling="2025-12-01 15:50:20.796664497 +0000 UTC m=+922.622410631" observedRunningTime="2025-12-01 15:50:22.771921189 +0000 UTC m=+924.597667303" watchObservedRunningTime="2025-12-01 15:50:22.774828028 +0000 UTC m=+924.600574122" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.810516 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-d2dxq" podStartSLOduration=3.93026778 podStartE2EDuration="33.810501504s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:51.044746025 +0000 UTC m=+892.870492109" lastFinishedPulling="2025-12-01 15:50:20.924979739 +0000 UTC m=+922.750725833" observedRunningTime="2025-12-01 15:50:22.809219555 +0000 UTC m=+924.634965649" watchObservedRunningTime="2025-12-01 15:50:22.810501504 +0000 UTC m=+924.636247598" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.846438 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc"] Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.909595 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-pqqzm" podStartSLOduration=4.104493744 podStartE2EDuration="33.909579348s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:51.000127354 +0000 UTC m=+892.825873458" lastFinishedPulling="2025-12-01 15:50:20.805212968 +0000 UTC m=+922.630959062" observedRunningTime="2025-12-01 15:50:22.908847335 +0000 UTC m=+924.734593419" watchObservedRunningTime="2025-12-01 15:50:22.909579348 +0000 UTC m=+924.735325442" Dec 01 15:50:22 crc kubenswrapper[4739]: I1201 15:50:22.974052 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xgpkt" podStartSLOduration=4.592392474 podStartE2EDuration="33.974033188s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:51.520663718 +0000 UTC m=+893.346409812" lastFinishedPulling="2025-12-01 15:50:20.902304432 +0000 UTC m=+922.728050526" observedRunningTime="2025-12-01 15:50:22.973512442 +0000 UTC m=+924.799258526" watchObservedRunningTime="2025-12-01 15:50:22.974033188 +0000 UTC m=+924.799779282" Dec 01 15:50:23 crc kubenswrapper[4739]: I1201 15:50:23.003165 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-px5zx" podStartSLOduration=4.546248259 podStartE2EDuration="34.003140772s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:51.442718816 +0000 UTC m=+893.268464910" lastFinishedPulling="2025-12-01 15:50:20.899611319 +0000 UTC m=+922.725357423" observedRunningTime="2025-12-01 15:50:22.998058466 +0000 UTC m=+924.823804560" watchObservedRunningTime="2025-12-01 15:50:23.003140772 +0000 UTC m=+924.828886866" Dec 01 15:50:23 crc kubenswrapper[4739]: I1201 15:50:23.687992 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" event={"ID":"5f474df5-ced8-45ff-840a-c52898696bf7","Type":"ContainerStarted","Data":"3ee374d21e79cb55cad7228746d8eded6dd84e5d339ca17ec00ad5ce524e93db"} Dec 01 15:50:23 crc kubenswrapper[4739]: I1201 15:50:23.688037 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" event={"ID":"5f474df5-ced8-45ff-840a-c52898696bf7","Type":"ContainerStarted","Data":"7378a0fd22ddafd40c4bf42666d5e9871340e6ef97c5cdc207c4cf67478f49b6"} Dec 01 15:50:23 crc kubenswrapper[4739]: I1201 15:50:23.694522 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fxvzt" Dec 01 15:50:23 crc kubenswrapper[4739]: I1201 15:50:23.694567 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4pwqm" Dec 01 15:50:23 crc kubenswrapper[4739]: I1201 15:50:23.783022 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" podStartSLOduration=33.783007151 podStartE2EDuration="33.783007151s" podCreationTimestamp="2025-12-01 15:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:50:23.773443567 +0000 UTC m=+925.599189671" watchObservedRunningTime="2025-12-01 15:50:23.783007151 +0000 UTC m=+925.608753245" Dec 01 15:50:24 crc kubenswrapper[4739]: I1201 15:50:24.697106 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:50:25 crc kubenswrapper[4739]: I1201 15:50:25.706952 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4nwgp" event={"ID":"903f2159-9ae0-4b78-8447-3b5bd4a420ef","Type":"ContainerStarted","Data":"77b3e8b3722b9b7abcbb4ce9afe9b34f0b6f8788c65603a0085dfbc3f5ea3fe8"} Dec 01 15:50:25 crc kubenswrapper[4739]: I1201 15:50:25.707211 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4nwgp" Dec 01 15:50:25 crc kubenswrapper[4739]: I1201 15:50:25.712646 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-2m26g" event={"ID":"8525c2bf-aedc-4816-8ea5-4ebdda160f3d","Type":"ContainerStarted","Data":"a35f722c1af08d86b86d08359d1e11ef011ca9d92b25cab880655060ffe8b055"} Dec 01 15:50:25 crc kubenswrapper[4739]: I1201 15:50:25.712833 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-2m26g" Dec 01 15:50:25 crc kubenswrapper[4739]: I1201 15:50:25.714671 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zrhx2" event={"ID":"c6ddb146-59dc-41c2-9278-466eda24296c","Type":"ContainerStarted","Data":"d22a826fdf01a428ac2871f6c47587edeb78d34f50395f80d2e1f1d1d65c6fb3"} Dec 01 15:50:25 crc kubenswrapper[4739]: I1201 15:50:25.714845 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zrhx2" Dec 01 15:50:25 crc kubenswrapper[4739]: I1201 15:50:25.730507 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4nwgp" podStartSLOduration=4.540488577 podStartE2EDuration="36.73048866s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:51.135532203 +0000 UTC m=+892.961278297" lastFinishedPulling="2025-12-01 15:50:23.325532286 +0000 UTC m=+925.151278380" observedRunningTime="2025-12-01 15:50:25.72398515 +0000 UTC m=+927.549731264" watchObservedRunningTime="2025-12-01 15:50:25.73048866 +0000 UTC m=+927.556234754" Dec 01 15:50:25 crc kubenswrapper[4739]: I1201 15:50:25.755069 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-2m26g" podStartSLOduration=4.429984812 podStartE2EDuration="36.755049594s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:50.998761902 +0000 UTC m=+892.824507996" lastFinishedPulling="2025-12-01 15:50:23.323826684 +0000 UTC m=+925.149572778" observedRunningTime="2025-12-01 15:50:25.747232654 +0000 UTC m=+927.572978788" watchObservedRunningTime="2025-12-01 15:50:25.755049594 +0000 UTC m=+927.580795688" Dec 01 15:50:25 crc kubenswrapper[4739]: I1201 15:50:25.765644 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zrhx2" podStartSLOduration=4.589886316 podStartE2EDuration="36.76562413s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:51.268245822 +0000 UTC m=+893.093991916" lastFinishedPulling="2025-12-01 15:50:23.443983636 +0000 UTC m=+925.269729730" observedRunningTime="2025-12-01 15:50:25.762881456 +0000 UTC m=+927.588627550" watchObservedRunningTime="2025-12-01 15:50:25.76562413 +0000 UTC m=+927.591370234" Dec 01 15:50:27 crc kubenswrapper[4739]: I1201 15:50:27.728704 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" event={"ID":"54dcd8b3-40f6-4487-bc4f-810c6aebb6fe","Type":"ContainerStarted","Data":"877705cc02f0b7e81a1c2b93a94e3f505c470690a72fc6de5ae4e2922ff4f419"} Dec 01 15:50:27 crc kubenswrapper[4739]: I1201 15:50:27.729066 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" event={"ID":"54dcd8b3-40f6-4487-bc4f-810c6aebb6fe","Type":"ContainerStarted","Data":"abc69a9101cad860f20809c2ecc00c99ac5631e73e16157a8e140077fa43bace"} Dec 01 15:50:27 crc kubenswrapper[4739]: I1201 15:50:27.729099 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:50:27 crc kubenswrapper[4739]: I1201 15:50:27.730607 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq" event={"ID":"28db1029-7abe-4d58-afbd-0022422c286c","Type":"ContainerStarted","Data":"0e177ed525afea300b9d37aa5e8fae34394c92dc3e7c02b25ea84859d9ddcdc1"} Dec 01 15:50:27 crc kubenswrapper[4739]: I1201 15:50:27.730856 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq" Dec 01 15:50:27 crc kubenswrapper[4739]: I1201 15:50:27.751907 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" podStartSLOduration=34.586260376 podStartE2EDuration="38.7518798s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:50:22.423751662 +0000 UTC m=+924.249497756" lastFinishedPulling="2025-12-01 15:50:26.589371086 +0000 UTC m=+928.415117180" observedRunningTime="2025-12-01 15:50:27.751529089 +0000 UTC m=+929.577275203" watchObservedRunningTime="2025-12-01 15:50:27.7518798 +0000 UTC m=+929.577625934" Dec 01 15:50:27 crc kubenswrapper[4739]: I1201 15:50:27.782171 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq" podStartSLOduration=2.912370816 podStartE2EDuration="38.78214855s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:50.706124288 +0000 UTC m=+892.531870382" lastFinishedPulling="2025-12-01 15:50:26.575902022 +0000 UTC m=+928.401648116" observedRunningTime="2025-12-01 15:50:27.770971856 +0000 UTC m=+929.596717950" watchObservedRunningTime="2025-12-01 15:50:27.78214855 +0000 UTC m=+929.607894644" Dec 01 15:50:30 crc kubenswrapper[4739]: I1201 15:50:30.134547 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-2m26g" Dec 01 15:50:30 crc kubenswrapper[4739]: I1201 15:50:30.325279 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-4nwgp" Dec 01 15:50:30 crc kubenswrapper[4739]: I1201 15:50:30.342486 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zrhx2" Dec 01 15:50:31 crc kubenswrapper[4739]: I1201 15:50:31.215880 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:50:31 crc kubenswrapper[4739]: I1201 15:50:31.215936 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:50:31 crc kubenswrapper[4739]: I1201 15:50:31.293258 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:50:31 crc kubenswrapper[4739]: I1201 15:50:31.784581 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gpkcn" Dec 01 15:50:31 crc kubenswrapper[4739]: I1201 15:50:31.873666 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:50:31 crc kubenswrapper[4739]: I1201 15:50:31.939582 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ll5cw"] Dec 01 15:50:32 crc kubenswrapper[4739]: I1201 15:50:32.346769 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-59fc5775d4-chxkc" Dec 01 15:50:33 crc kubenswrapper[4739]: I1201 15:50:33.480035 4739 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 15:50:33 crc kubenswrapper[4739]: I1201 15:50:33.793655 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ll5cw" podUID="41c3fd88-c7a2-4741-a3da-6b9f33b42b46" containerName="registry-server" containerID="cri-o://35cc6cc2101ab475c7d8b59b8b43719f372c6d7020913f2e849277da0f049a3e" gracePeriod=2 Dec 01 15:50:35 crc kubenswrapper[4739]: I1201 15:50:35.915363 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh" Dec 01 15:50:36 crc kubenswrapper[4739]: I1201 15:50:36.835881 4739 generic.go:334] "Generic (PLEG): container finished" podID="41c3fd88-c7a2-4741-a3da-6b9f33b42b46" containerID="35cc6cc2101ab475c7d8b59b8b43719f372c6d7020913f2e849277da0f049a3e" exitCode=0 Dec 01 15:50:36 crc kubenswrapper[4739]: I1201 15:50:36.835969 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ll5cw" event={"ID":"41c3fd88-c7a2-4741-a3da-6b9f33b42b46","Type":"ContainerDied","Data":"35cc6cc2101ab475c7d8b59b8b43719f372c6d7020913f2e849277da0f049a3e"} Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.139640 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.244480 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mzzj\" (UniqueName: \"kubernetes.io/projected/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-kube-api-access-6mzzj\") pod \"41c3fd88-c7a2-4741-a3da-6b9f33b42b46\" (UID: \"41c3fd88-c7a2-4741-a3da-6b9f33b42b46\") " Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.244571 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-utilities\") pod \"41c3fd88-c7a2-4741-a3da-6b9f33b42b46\" (UID: \"41c3fd88-c7a2-4741-a3da-6b9f33b42b46\") " Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.244594 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-catalog-content\") pod \"41c3fd88-c7a2-4741-a3da-6b9f33b42b46\" (UID: \"41c3fd88-c7a2-4741-a3da-6b9f33b42b46\") " Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.245761 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-utilities" (OuterVolumeSpecName: "utilities") pod "41c3fd88-c7a2-4741-a3da-6b9f33b42b46" (UID: "41c3fd88-c7a2-4741-a3da-6b9f33b42b46"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.250245 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-kube-api-access-6mzzj" (OuterVolumeSpecName: "kube-api-access-6mzzj") pod "41c3fd88-c7a2-4741-a3da-6b9f33b42b46" (UID: "41c3fd88-c7a2-4741-a3da-6b9f33b42b46"). InnerVolumeSpecName "kube-api-access-6mzzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.298573 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41c3fd88-c7a2-4741-a3da-6b9f33b42b46" (UID: "41c3fd88-c7a2-4741-a3da-6b9f33b42b46"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.346031 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mzzj\" (UniqueName: \"kubernetes.io/projected/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-kube-api-access-6mzzj\") on node \"crc\" DevicePath \"\"" Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.346063 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.346072 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c3fd88-c7a2-4741-a3da-6b9f33b42b46-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.845187 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd" event={"ID":"8f8c53db-885f-4548-bcc0-02ce36cd6543","Type":"ContainerStarted","Data":"25670179e41fabb4fd32e6a9b7ec32a5f4f91dda43f9280d8fc03f2b9a5ef308"} Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.845468 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd" Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.847544 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ll5cw" event={"ID":"41c3fd88-c7a2-4741-a3da-6b9f33b42b46","Type":"ContainerDied","Data":"bf439f4ca6554eec30c5e8667a4ce3724a283bb03fa2a9a700644f966164066f"} Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.847641 4739 scope.go:117] "RemoveContainer" containerID="35cc6cc2101ab475c7d8b59b8b43719f372c6d7020913f2e849277da0f049a3e" Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.847577 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ll5cw" Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.849514 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt" event={"ID":"974d1918-6e6a-4660-898c-b8d8db439f46","Type":"ContainerStarted","Data":"83ae9464097269a57ff7b9e76b827805eefb186e2fc167fddbd60e9d8f7d28b3"} Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.850104 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt" Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.867063 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd" podStartSLOduration=2.903667202 podStartE2EDuration="48.867030882s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:51.044662292 +0000 UTC m=+892.870408386" lastFinishedPulling="2025-12-01 15:50:37.008025972 +0000 UTC m=+938.833772066" observedRunningTime="2025-12-01 15:50:37.863051029 +0000 UTC m=+939.688797163" watchObservedRunningTime="2025-12-01 15:50:37.867030882 +0000 UTC m=+939.692777016" Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.868729 4739 scope.go:117] "RemoveContainer" containerID="fe05f1a56b1e6cf55406a435c84d8609e7fa5e5e6855a05aeda63fdc6a68f1c9" Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.888631 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt" podStartSLOduration=3.191642327 podStartE2EDuration="48.888609834s" podCreationTimestamp="2025-12-01 15:49:49 +0000 UTC" firstStartedPulling="2025-12-01 15:49:51.121059148 +0000 UTC m=+892.946805242" lastFinishedPulling="2025-12-01 15:50:36.818026655 +0000 UTC m=+938.643772749" observedRunningTime="2025-12-01 15:50:37.886474239 +0000 UTC m=+939.712220333" watchObservedRunningTime="2025-12-01 15:50:37.888609834 +0000 UTC m=+939.714355968" Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.894914 4739 scope.go:117] "RemoveContainer" containerID="bbae966eb5cf115e28348fd122e3a0921f0cd1e5dc9ae0811af2579fe7c3104a" Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.913762 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ll5cw"] Dec 01 15:50:37 crc kubenswrapper[4739]: I1201 15:50:37.921349 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ll5cw"] Dec 01 15:50:38 crc kubenswrapper[4739]: I1201 15:50:38.489680 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41c3fd88-c7a2-4741-a3da-6b9f33b42b46" path="/var/lib/kubelet/pods/41c3fd88-c7a2-4741-a3da-6b9f33b42b46/volumes" Dec 01 15:50:39 crc kubenswrapper[4739]: I1201 15:50:39.622660 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:50:39 crc kubenswrapper[4739]: I1201 15:50:39.622714 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:50:39 crc kubenswrapper[4739]: I1201 15:50:39.622766 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:50:39 crc kubenswrapper[4739]: I1201 15:50:39.623195 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c739ec8ed0d32f48e79459013f2a8032f453df5dc42b21acdeda98c2b463c62e"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:50:39 crc kubenswrapper[4739]: I1201 15:50:39.623246 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://c739ec8ed0d32f48e79459013f2a8032f453df5dc42b21acdeda98c2b463c62e" gracePeriod=600 Dec 01 15:50:39 crc kubenswrapper[4739]: I1201 15:50:39.764070 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-5d986d85c9-2s7sq" Dec 01 15:50:39 crc kubenswrapper[4739]: I1201 15:50:39.872070 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="c739ec8ed0d32f48e79459013f2a8032f453df5dc42b21acdeda98c2b463c62e" exitCode=0 Dec 01 15:50:39 crc kubenswrapper[4739]: I1201 15:50:39.872103 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"c739ec8ed0d32f48e79459013f2a8032f453df5dc42b21acdeda98c2b463c62e"} Dec 01 15:50:39 crc kubenswrapper[4739]: I1201 15:50:39.872490 4739 scope.go:117] "RemoveContainer" containerID="e3830d5cb4e555123caab94e2aaa75fd65f71b1d428fbec87bdc6d807609b113" Dec 01 15:50:40 crc kubenswrapper[4739]: I1201 15:50:40.884937 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"d23f235659cb740130c9b9ee149b02d7de5733c24075029cc9884583e5415dad"} Dec 01 15:50:49 crc kubenswrapper[4739]: I1201 15:50:49.988057 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-758wd" Dec 01 15:50:50 crc kubenswrapper[4739]: I1201 15:50:50.248036 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-8bsmt" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.301239 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-499zj"] Dec 01 15:51:05 crc kubenswrapper[4739]: E1201 15:51:05.302213 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41c3fd88-c7a2-4741-a3da-6b9f33b42b46" containerName="extract-utilities" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.302229 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="41c3fd88-c7a2-4741-a3da-6b9f33b42b46" containerName="extract-utilities" Dec 01 15:51:05 crc kubenswrapper[4739]: E1201 15:51:05.302275 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41c3fd88-c7a2-4741-a3da-6b9f33b42b46" containerName="registry-server" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.302283 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="41c3fd88-c7a2-4741-a3da-6b9f33b42b46" containerName="registry-server" Dec 01 15:51:05 crc kubenswrapper[4739]: E1201 15:51:05.302306 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41c3fd88-c7a2-4741-a3da-6b9f33b42b46" containerName="extract-content" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.302314 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="41c3fd88-c7a2-4741-a3da-6b9f33b42b46" containerName="extract-content" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.302531 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="41c3fd88-c7a2-4741-a3da-6b9f33b42b46" containerName="registry-server" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.303403 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-499zj" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.306556 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-499zj"] Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.306982 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.306987 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.307001 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.307891 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-kj768" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.349166 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h4ntd"] Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.350648 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.355007 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.370747 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h4ntd"] Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.378930 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09152c95-880e-45e7-8ba8-d247d5c5cfc3-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-h4ntd\" (UID: \"09152c95-880e-45e7-8ba8-d247d5c5cfc3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.379013 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09152c95-880e-45e7-8ba8-d247d5c5cfc3-config\") pod \"dnsmasq-dns-78dd6ddcc-h4ntd\" (UID: \"09152c95-880e-45e7-8ba8-d247d5c5cfc3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.379046 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7kxb\" (UniqueName: \"kubernetes.io/projected/09152c95-880e-45e7-8ba8-d247d5c5cfc3-kube-api-access-b7kxb\") pod \"dnsmasq-dns-78dd6ddcc-h4ntd\" (UID: \"09152c95-880e-45e7-8ba8-d247d5c5cfc3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.379072 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68e76809-d9d8-4d93-8213-cde5d7eb5852-config\") pod \"dnsmasq-dns-675f4bcbfc-499zj\" (UID: \"68e76809-d9d8-4d93-8213-cde5d7eb5852\") " pod="openstack/dnsmasq-dns-675f4bcbfc-499zj" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.379095 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m57q\" (UniqueName: \"kubernetes.io/projected/68e76809-d9d8-4d93-8213-cde5d7eb5852-kube-api-access-4m57q\") pod \"dnsmasq-dns-675f4bcbfc-499zj\" (UID: \"68e76809-d9d8-4d93-8213-cde5d7eb5852\") " pod="openstack/dnsmasq-dns-675f4bcbfc-499zj" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.480352 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09152c95-880e-45e7-8ba8-d247d5c5cfc3-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-h4ntd\" (UID: \"09152c95-880e-45e7-8ba8-d247d5c5cfc3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.480442 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09152c95-880e-45e7-8ba8-d247d5c5cfc3-config\") pod \"dnsmasq-dns-78dd6ddcc-h4ntd\" (UID: \"09152c95-880e-45e7-8ba8-d247d5c5cfc3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.480480 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7kxb\" (UniqueName: \"kubernetes.io/projected/09152c95-880e-45e7-8ba8-d247d5c5cfc3-kube-api-access-b7kxb\") pod \"dnsmasq-dns-78dd6ddcc-h4ntd\" (UID: \"09152c95-880e-45e7-8ba8-d247d5c5cfc3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.480508 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68e76809-d9d8-4d93-8213-cde5d7eb5852-config\") pod \"dnsmasq-dns-675f4bcbfc-499zj\" (UID: \"68e76809-d9d8-4d93-8213-cde5d7eb5852\") " pod="openstack/dnsmasq-dns-675f4bcbfc-499zj" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.480531 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m57q\" (UniqueName: \"kubernetes.io/projected/68e76809-d9d8-4d93-8213-cde5d7eb5852-kube-api-access-4m57q\") pod \"dnsmasq-dns-675f4bcbfc-499zj\" (UID: \"68e76809-d9d8-4d93-8213-cde5d7eb5852\") " pod="openstack/dnsmasq-dns-675f4bcbfc-499zj" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.481245 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09152c95-880e-45e7-8ba8-d247d5c5cfc3-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-h4ntd\" (UID: \"09152c95-880e-45e7-8ba8-d247d5c5cfc3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.481540 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09152c95-880e-45e7-8ba8-d247d5c5cfc3-config\") pod \"dnsmasq-dns-78dd6ddcc-h4ntd\" (UID: \"09152c95-880e-45e7-8ba8-d247d5c5cfc3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.481744 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68e76809-d9d8-4d93-8213-cde5d7eb5852-config\") pod \"dnsmasq-dns-675f4bcbfc-499zj\" (UID: \"68e76809-d9d8-4d93-8213-cde5d7eb5852\") " pod="openstack/dnsmasq-dns-675f4bcbfc-499zj" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.504551 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m57q\" (UniqueName: \"kubernetes.io/projected/68e76809-d9d8-4d93-8213-cde5d7eb5852-kube-api-access-4m57q\") pod \"dnsmasq-dns-675f4bcbfc-499zj\" (UID: \"68e76809-d9d8-4d93-8213-cde5d7eb5852\") " pod="openstack/dnsmasq-dns-675f4bcbfc-499zj" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.507040 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7kxb\" (UniqueName: \"kubernetes.io/projected/09152c95-880e-45e7-8ba8-d247d5c5cfc3-kube-api-access-b7kxb\") pod \"dnsmasq-dns-78dd6ddcc-h4ntd\" (UID: \"09152c95-880e-45e7-8ba8-d247d5c5cfc3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.625910 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-499zj" Dec 01 15:51:05 crc kubenswrapper[4739]: I1201 15:51:05.671169 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" Dec 01 15:51:06 crc kubenswrapper[4739]: I1201 15:51:06.113514 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-499zj"] Dec 01 15:51:06 crc kubenswrapper[4739]: I1201 15:51:06.127111 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-499zj" event={"ID":"68e76809-d9d8-4d93-8213-cde5d7eb5852","Type":"ContainerStarted","Data":"1424f44cb82bb0afe6a5c115db913a495b78ab18e897773b3c295feb14fc2f12"} Dec 01 15:51:06 crc kubenswrapper[4739]: I1201 15:51:06.172111 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h4ntd"] Dec 01 15:51:06 crc kubenswrapper[4739]: W1201 15:51:06.175601 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09152c95_880e_45e7_8ba8_d247d5c5cfc3.slice/crio-444e239232841adb91349c692d911fb4c78b4fac360dab2df2240da16a43c8ca WatchSource:0}: Error finding container 444e239232841adb91349c692d911fb4c78b4fac360dab2df2240da16a43c8ca: Status 404 returned error can't find the container with id 444e239232841adb91349c692d911fb4c78b4fac360dab2df2240da16a43c8ca Dec 01 15:51:07 crc kubenswrapper[4739]: I1201 15:51:07.134026 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" event={"ID":"09152c95-880e-45e7-8ba8-d247d5c5cfc3","Type":"ContainerStarted","Data":"444e239232841adb91349c692d911fb4c78b4fac360dab2df2240da16a43c8ca"} Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.369765 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-499zj"] Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.386353 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7hv4f"] Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.393028 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.402248 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7hv4f"] Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.429673 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbg58\" (UniqueName: \"kubernetes.io/projected/6470b09d-b153-4b28-96d4-9425fbbf9e18-kube-api-access-bbg58\") pod \"dnsmasq-dns-666b6646f7-7hv4f\" (UID: \"6470b09d-b153-4b28-96d4-9425fbbf9e18\") " pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.429732 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6470b09d-b153-4b28-96d4-9425fbbf9e18-config\") pod \"dnsmasq-dns-666b6646f7-7hv4f\" (UID: \"6470b09d-b153-4b28-96d4-9425fbbf9e18\") " pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.429791 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6470b09d-b153-4b28-96d4-9425fbbf9e18-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7hv4f\" (UID: \"6470b09d-b153-4b28-96d4-9425fbbf9e18\") " pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.531879 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6470b09d-b153-4b28-96d4-9425fbbf9e18-config\") pod \"dnsmasq-dns-666b6646f7-7hv4f\" (UID: \"6470b09d-b153-4b28-96d4-9425fbbf9e18\") " pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.534756 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6470b09d-b153-4b28-96d4-9425fbbf9e18-config\") pod \"dnsmasq-dns-666b6646f7-7hv4f\" (UID: \"6470b09d-b153-4b28-96d4-9425fbbf9e18\") " pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.534999 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6470b09d-b153-4b28-96d4-9425fbbf9e18-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7hv4f\" (UID: \"6470b09d-b153-4b28-96d4-9425fbbf9e18\") " pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.535185 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbg58\" (UniqueName: \"kubernetes.io/projected/6470b09d-b153-4b28-96d4-9425fbbf9e18-kube-api-access-bbg58\") pod \"dnsmasq-dns-666b6646f7-7hv4f\" (UID: \"6470b09d-b153-4b28-96d4-9425fbbf9e18\") " pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.536966 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6470b09d-b153-4b28-96d4-9425fbbf9e18-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7hv4f\" (UID: \"6470b09d-b153-4b28-96d4-9425fbbf9e18\") " pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.559531 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbg58\" (UniqueName: \"kubernetes.io/projected/6470b09d-b153-4b28-96d4-9425fbbf9e18-kube-api-access-bbg58\") pod \"dnsmasq-dns-666b6646f7-7hv4f\" (UID: \"6470b09d-b153-4b28-96d4-9425fbbf9e18\") " pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.663616 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h4ntd"] Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.684884 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jr7xb"] Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.688593 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.699818 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jr7xb"] Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.711459 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.739958 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ccf3a0a-3522-4825-9a2b-b926a404e642-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-jr7xb\" (UID: \"5ccf3a0a-3522-4825-9a2b-b926a404e642\") " pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.740040 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ccf3a0a-3522-4825-9a2b-b926a404e642-config\") pod \"dnsmasq-dns-57d769cc4f-jr7xb\" (UID: \"5ccf3a0a-3522-4825-9a2b-b926a404e642\") " pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.740063 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvgn6\" (UniqueName: \"kubernetes.io/projected/5ccf3a0a-3522-4825-9a2b-b926a404e642-kube-api-access-cvgn6\") pod \"dnsmasq-dns-57d769cc4f-jr7xb\" (UID: \"5ccf3a0a-3522-4825-9a2b-b926a404e642\") " pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.840776 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ccf3a0a-3522-4825-9a2b-b926a404e642-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-jr7xb\" (UID: \"5ccf3a0a-3522-4825-9a2b-b926a404e642\") " pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.840855 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ccf3a0a-3522-4825-9a2b-b926a404e642-config\") pod \"dnsmasq-dns-57d769cc4f-jr7xb\" (UID: \"5ccf3a0a-3522-4825-9a2b-b926a404e642\") " pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.840879 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvgn6\" (UniqueName: \"kubernetes.io/projected/5ccf3a0a-3522-4825-9a2b-b926a404e642-kube-api-access-cvgn6\") pod \"dnsmasq-dns-57d769cc4f-jr7xb\" (UID: \"5ccf3a0a-3522-4825-9a2b-b926a404e642\") " pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.842112 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ccf3a0a-3522-4825-9a2b-b926a404e642-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-jr7xb\" (UID: \"5ccf3a0a-3522-4825-9a2b-b926a404e642\") " pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.842662 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ccf3a0a-3522-4825-9a2b-b926a404e642-config\") pod \"dnsmasq-dns-57d769cc4f-jr7xb\" (UID: \"5ccf3a0a-3522-4825-9a2b-b926a404e642\") " pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" Dec 01 15:51:08 crc kubenswrapper[4739]: I1201 15:51:08.875383 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvgn6\" (UniqueName: \"kubernetes.io/projected/5ccf3a0a-3522-4825-9a2b-b926a404e642-kube-api-access-cvgn6\") pod \"dnsmasq-dns-57d769cc4f-jr7xb\" (UID: \"5ccf3a0a-3522-4825-9a2b-b926a404e642\") " pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.016760 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.336225 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jr7xb"] Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.356973 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7hv4f"] Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.540687 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.543098 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.548815 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.549053 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.549264 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.549469 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.549724 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-h8p42" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.549943 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.550156 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.563730 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.663162 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.663212 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.663273 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.663313 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.663470 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.663589 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.663656 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f128863-acb7-4080-829a-01bd78766374-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.663688 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.663773 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-config-data\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.663830 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvk9f\" (UniqueName: \"kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-kube-api-access-fvk9f\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.663887 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f128863-acb7-4080-829a-01bd78766374-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.765008 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.765066 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.765114 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.765156 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.765209 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f128863-acb7-4080-829a-01bd78766374-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.765250 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.765295 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-config-data\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.765323 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvk9f\" (UniqueName: \"kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-kube-api-access-fvk9f\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.765356 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f128863-acb7-4080-829a-01bd78766374-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.765392 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.765433 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.766455 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.766526 4739 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.766849 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.766893 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.767458 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-config-data\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.776987 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.777322 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f128863-acb7-4080-829a-01bd78766374-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.785946 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.787101 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.788089 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvk9f\" (UniqueName: \"kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-kube-api-access-fvk9f\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.791531 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f128863-acb7-4080-829a-01bd78766374-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.833113 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.837449 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.839647 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.846943 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-x2bjw" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.847227 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.847397 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.847551 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.847792 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.847965 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.848116 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.850760 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.868567 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.868624 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.868665 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.868696 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6fq8\" (UniqueName: \"kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-kube-api-access-s6fq8\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.868745 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.868760 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.868776 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2606100-66af-4acd-98a5-2edcfb281544-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.868794 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.868810 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.868830 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2606100-66af-4acd-98a5-2edcfb281544-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.868854 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.892620 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.970602 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.970653 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.970674 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.970709 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.970739 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6fq8\" (UniqueName: \"kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-kube-api-access-s6fq8\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.970790 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.970806 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.970842 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2606100-66af-4acd-98a5-2edcfb281544-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.970860 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.970877 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.970894 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2606100-66af-4acd-98a5-2edcfb281544-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.971393 4739 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.971685 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.971715 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.972069 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.972710 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.976056 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.976083 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.976441 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2606100-66af-4acd-98a5-2edcfb281544-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.988792 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.989616 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2606100-66af-4acd-98a5-2edcfb281544-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:09 crc kubenswrapper[4739]: I1201 15:51:09.989969 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6fq8\" (UniqueName: \"kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-kube-api-access-s6fq8\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:10 crc kubenswrapper[4739]: I1201 15:51:10.005335 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:10 crc kubenswrapper[4739]: I1201 15:51:10.192183 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.169742 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.172631 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.175116 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-59vfb" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.175401 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.175740 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.175811 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.180926 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.184277 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.296265 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8227dcb8-9b5f-4348-bba7-1560fa84861a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.296316 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8227dcb8-9b5f-4348-bba7-1560fa84861a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.296340 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8227dcb8-9b5f-4348-bba7-1560fa84861a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.296385 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8227dcb8-9b5f-4348-bba7-1560fa84861a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.296781 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.296932 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8227dcb8-9b5f-4348-bba7-1560fa84861a-config-data-default\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.297002 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd5b8\" (UniqueName: \"kubernetes.io/projected/8227dcb8-9b5f-4348-bba7-1560fa84861a-kube-api-access-sd5b8\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.297097 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8227dcb8-9b5f-4348-bba7-1560fa84861a-kolla-config\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.399200 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8227dcb8-9b5f-4348-bba7-1560fa84861a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.399263 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8227dcb8-9b5f-4348-bba7-1560fa84861a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.399291 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8227dcb8-9b5f-4348-bba7-1560fa84861a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.399342 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8227dcb8-9b5f-4348-bba7-1560fa84861a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.399380 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.399428 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8227dcb8-9b5f-4348-bba7-1560fa84861a-config-data-default\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.399457 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd5b8\" (UniqueName: \"kubernetes.io/projected/8227dcb8-9b5f-4348-bba7-1560fa84861a-kube-api-access-sd5b8\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.399494 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8227dcb8-9b5f-4348-bba7-1560fa84861a-kolla-config\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.399848 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8227dcb8-9b5f-4348-bba7-1560fa84861a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.400217 4739 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.401249 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8227dcb8-9b5f-4348-bba7-1560fa84861a-config-data-default\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.403545 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8227dcb8-9b5f-4348-bba7-1560fa84861a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.406116 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8227dcb8-9b5f-4348-bba7-1560fa84861a-kolla-config\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.406546 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8227dcb8-9b5f-4348-bba7-1560fa84861a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.407017 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8227dcb8-9b5f-4348-bba7-1560fa84861a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.423960 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd5b8\" (UniqueName: \"kubernetes.io/projected/8227dcb8-9b5f-4348-bba7-1560fa84861a-kube-api-access-sd5b8\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.424037 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"8227dcb8-9b5f-4348-bba7-1560fa84861a\") " pod="openstack/openstack-galera-0" Dec 01 15:51:11 crc kubenswrapper[4739]: I1201 15:51:11.513123 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.750116 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.751982 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.757485 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-jqx87" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.757687 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.757928 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.758053 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.758045 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.926881 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktvht\" (UniqueName: \"kubernetes.io/projected/a5323c11-dc67-47e7-83be-4b8a52f4ebef-kube-api-access-ktvht\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.926941 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a5323c11-dc67-47e7-83be-4b8a52f4ebef-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.926971 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5323c11-dc67-47e7-83be-4b8a52f4ebef-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.927077 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.927099 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a5323c11-dc67-47e7-83be-4b8a52f4ebef-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.927119 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5323c11-dc67-47e7-83be-4b8a52f4ebef-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.927154 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a5323c11-dc67-47e7-83be-4b8a52f4ebef-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.927177 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5323c11-dc67-47e7-83be-4b8a52f4ebef-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.964697 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.965632 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.967979 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.968279 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.973388 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-hn2g5" Dec 01 15:51:12 crc kubenswrapper[4739]: I1201 15:51:12.982271 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.028909 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktvht\" (UniqueName: \"kubernetes.io/projected/a5323c11-dc67-47e7-83be-4b8a52f4ebef-kube-api-access-ktvht\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.028965 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a5323c11-dc67-47e7-83be-4b8a52f4ebef-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.028988 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5323c11-dc67-47e7-83be-4b8a52f4ebef-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.029046 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.029066 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a5323c11-dc67-47e7-83be-4b8a52f4ebef-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.029085 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5323c11-dc67-47e7-83be-4b8a52f4ebef-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.029116 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a5323c11-dc67-47e7-83be-4b8a52f4ebef-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.029139 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5323c11-dc67-47e7-83be-4b8a52f4ebef-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.029598 4739 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.031108 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a5323c11-dc67-47e7-83be-4b8a52f4ebef-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.031249 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5323c11-dc67-47e7-83be-4b8a52f4ebef-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.032157 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a5323c11-dc67-47e7-83be-4b8a52f4ebef-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.032388 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a5323c11-dc67-47e7-83be-4b8a52f4ebef-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.041094 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5323c11-dc67-47e7-83be-4b8a52f4ebef-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.041166 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5323c11-dc67-47e7-83be-4b8a52f4ebef-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.054273 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.054808 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktvht\" (UniqueName: \"kubernetes.io/projected/a5323c11-dc67-47e7-83be-4b8a52f4ebef-kube-api-access-ktvht\") pod \"openstack-cell1-galera-0\" (UID: \"a5323c11-dc67-47e7-83be-4b8a52f4ebef\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.093009 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.130606 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87h6t\" (UniqueName: \"kubernetes.io/projected/7301c45d-f6e6-4d47-b75d-d381033edb1d-kube-api-access-87h6t\") pod \"memcached-0\" (UID: \"7301c45d-f6e6-4d47-b75d-d381033edb1d\") " pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.130763 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7301c45d-f6e6-4d47-b75d-d381033edb1d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"7301c45d-f6e6-4d47-b75d-d381033edb1d\") " pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.130859 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7301c45d-f6e6-4d47-b75d-d381033edb1d-kolla-config\") pod \"memcached-0\" (UID: \"7301c45d-f6e6-4d47-b75d-d381033edb1d\") " pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.130979 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7301c45d-f6e6-4d47-b75d-d381033edb1d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"7301c45d-f6e6-4d47-b75d-d381033edb1d\") " pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.131030 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7301c45d-f6e6-4d47-b75d-d381033edb1d-config-data\") pod \"memcached-0\" (UID: \"7301c45d-f6e6-4d47-b75d-d381033edb1d\") " pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.232197 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87h6t\" (UniqueName: \"kubernetes.io/projected/7301c45d-f6e6-4d47-b75d-d381033edb1d-kube-api-access-87h6t\") pod \"memcached-0\" (UID: \"7301c45d-f6e6-4d47-b75d-d381033edb1d\") " pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.232285 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7301c45d-f6e6-4d47-b75d-d381033edb1d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"7301c45d-f6e6-4d47-b75d-d381033edb1d\") " pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.232328 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7301c45d-f6e6-4d47-b75d-d381033edb1d-kolla-config\") pod \"memcached-0\" (UID: \"7301c45d-f6e6-4d47-b75d-d381033edb1d\") " pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.232381 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7301c45d-f6e6-4d47-b75d-d381033edb1d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"7301c45d-f6e6-4d47-b75d-d381033edb1d\") " pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.232406 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7301c45d-f6e6-4d47-b75d-d381033edb1d-config-data\") pod \"memcached-0\" (UID: \"7301c45d-f6e6-4d47-b75d-d381033edb1d\") " pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.233450 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7301c45d-f6e6-4d47-b75d-d381033edb1d-config-data\") pod \"memcached-0\" (UID: \"7301c45d-f6e6-4d47-b75d-d381033edb1d\") " pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.233461 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7301c45d-f6e6-4d47-b75d-d381033edb1d-kolla-config\") pod \"memcached-0\" (UID: \"7301c45d-f6e6-4d47-b75d-d381033edb1d\") " pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.238002 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7301c45d-f6e6-4d47-b75d-d381033edb1d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"7301c45d-f6e6-4d47-b75d-d381033edb1d\") " pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.252630 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87h6t\" (UniqueName: \"kubernetes.io/projected/7301c45d-f6e6-4d47-b75d-d381033edb1d-kube-api-access-87h6t\") pod \"memcached-0\" (UID: \"7301c45d-f6e6-4d47-b75d-d381033edb1d\") " pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.252860 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7301c45d-f6e6-4d47-b75d-d381033edb1d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"7301c45d-f6e6-4d47-b75d-d381033edb1d\") " pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: I1201 15:51:13.296304 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 15:51:13 crc kubenswrapper[4739]: W1201 15:51:13.349944 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6470b09d_b153_4b28_96d4_9425fbbf9e18.slice/crio-d629e6e613ec25d499029a7849871a0ea706363a73d8f9f758e4fb0b9a9c1e1a WatchSource:0}: Error finding container d629e6e613ec25d499029a7849871a0ea706363a73d8f9f758e4fb0b9a9c1e1a: Status 404 returned error can't find the container with id d629e6e613ec25d499029a7849871a0ea706363a73d8f9f758e4fb0b9a9c1e1a Dec 01 15:51:14 crc kubenswrapper[4739]: I1201 15:51:14.203007 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" event={"ID":"6470b09d-b153-4b28-96d4-9425fbbf9e18","Type":"ContainerStarted","Data":"d629e6e613ec25d499029a7849871a0ea706363a73d8f9f758e4fb0b9a9c1e1a"} Dec 01 15:51:14 crc kubenswrapper[4739]: I1201 15:51:14.203950 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" event={"ID":"5ccf3a0a-3522-4825-9a2b-b926a404e642","Type":"ContainerStarted","Data":"d187bad60c30604096f9d6b070ef1f7132777bce3ebb289f6b1f96c7b90067ea"} Dec 01 15:51:14 crc kubenswrapper[4739]: I1201 15:51:14.862984 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 15:51:14 crc kubenswrapper[4739]: I1201 15:51:14.864489 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 15:51:14 crc kubenswrapper[4739]: I1201 15:51:14.868474 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-lwgjk" Dec 01 15:51:14 crc kubenswrapper[4739]: I1201 15:51:14.882805 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 15:51:14 crc kubenswrapper[4739]: I1201 15:51:14.966102 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldx8c\" (UniqueName: \"kubernetes.io/projected/51f0a0d6-a82c-495d-9cd4-141c824e6453-kube-api-access-ldx8c\") pod \"kube-state-metrics-0\" (UID: \"51f0a0d6-a82c-495d-9cd4-141c824e6453\") " pod="openstack/kube-state-metrics-0" Dec 01 15:51:15 crc kubenswrapper[4739]: I1201 15:51:15.067731 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldx8c\" (UniqueName: \"kubernetes.io/projected/51f0a0d6-a82c-495d-9cd4-141c824e6453-kube-api-access-ldx8c\") pod \"kube-state-metrics-0\" (UID: \"51f0a0d6-a82c-495d-9cd4-141c824e6453\") " pod="openstack/kube-state-metrics-0" Dec 01 15:51:15 crc kubenswrapper[4739]: I1201 15:51:15.100141 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldx8c\" (UniqueName: \"kubernetes.io/projected/51f0a0d6-a82c-495d-9cd4-141c824e6453-kube-api-access-ldx8c\") pod \"kube-state-metrics-0\" (UID: \"51f0a0d6-a82c-495d-9cd4-141c824e6453\") " pod="openstack/kube-state-metrics-0" Dec 01 15:51:15 crc kubenswrapper[4739]: I1201 15:51:15.271320 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.300537 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4ltrp"] Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.302266 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.306047 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.306080 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-4zksb" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.306699 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.321587 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-pr8vl"] Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.323494 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.333191 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ltrp"] Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.349288 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-pr8vl"] Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.357648 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbg8s\" (UniqueName: \"kubernetes.io/projected/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-kube-api-access-cbg8s\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.357711 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-var-log-ovn\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.357737 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0bed2a41-e961-4361-9316-2f29b6c0b0b5-var-run\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.357769 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-var-run\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.357794 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0bed2a41-e961-4361-9316-2f29b6c0b0b5-var-lib\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.357823 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zjzl\" (UniqueName: \"kubernetes.io/projected/0bed2a41-e961-4361-9316-2f29b6c0b0b5-kube-api-access-7zjzl\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.357860 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-scripts\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.357884 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-ovn-controller-tls-certs\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.357916 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-combined-ca-bundle\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.357943 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0bed2a41-e961-4361-9316-2f29b6c0b0b5-etc-ovs\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.357965 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0bed2a41-e961-4361-9316-2f29b6c0b0b5-var-log\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.357984 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0bed2a41-e961-4361-9316-2f29b6c0b0b5-scripts\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.358022 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-var-run-ovn\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459210 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-var-run\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459271 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0bed2a41-e961-4361-9316-2f29b6c0b0b5-var-lib\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459321 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zjzl\" (UniqueName: \"kubernetes.io/projected/0bed2a41-e961-4361-9316-2f29b6c0b0b5-kube-api-access-7zjzl\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459361 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-scripts\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459386 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-ovn-controller-tls-certs\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459438 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-combined-ca-bundle\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459467 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0bed2a41-e961-4361-9316-2f29b6c0b0b5-etc-ovs\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459493 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0bed2a41-e961-4361-9316-2f29b6c0b0b5-var-log\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459514 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0bed2a41-e961-4361-9316-2f29b6c0b0b5-scripts\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459556 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-var-run-ovn\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459590 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbg8s\" (UniqueName: \"kubernetes.io/projected/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-kube-api-access-cbg8s\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459629 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-var-log-ovn\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459653 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0bed2a41-e961-4361-9316-2f29b6c0b0b5-var-run\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459727 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0bed2a41-e961-4361-9316-2f29b6c0b0b5-var-lib\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459805 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0bed2a41-e961-4361-9316-2f29b6c0b0b5-var-run\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459815 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-var-run\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.459861 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0bed2a41-e961-4361-9316-2f29b6c0b0b5-etc-ovs\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.460019 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0bed2a41-e961-4361-9316-2f29b6c0b0b5-var-log\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.460449 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-var-run-ovn\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.461616 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-var-log-ovn\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.461833 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-scripts\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.462246 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0bed2a41-e961-4361-9316-2f29b6c0b0b5-scripts\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.466252 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-ovn-controller-tls-certs\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.472894 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-combined-ca-bundle\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.475231 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbg8s\" (UniqueName: \"kubernetes.io/projected/10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15-kube-api-access-cbg8s\") pod \"ovn-controller-4ltrp\" (UID: \"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15\") " pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.479895 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zjzl\" (UniqueName: \"kubernetes.io/projected/0bed2a41-e961-4361-9316-2f29b6c0b0b5-kube-api-access-7zjzl\") pod \"ovn-controller-ovs-pr8vl\" (UID: \"0bed2a41-e961-4361-9316-2f29b6c0b0b5\") " pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.642535 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:19 crc kubenswrapper[4739]: I1201 15:51:19.656185 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.720263 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.721915 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.723919 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.724107 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.724113 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.724671 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.724864 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-gvvvf" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.728251 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.872290 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.873771 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.877773 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.877926 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.877862 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.877948 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-n6bvj" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.885023 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.900898 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.900986 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17dd4503-6f0a-48f1-b661-d78154c9a05d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.901029 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17dd4503-6f0a-48f1-b661-d78154c9a05d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.901049 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/17dd4503-6f0a-48f1-b661-d78154c9a05d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.901084 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/17dd4503-6f0a-48f1-b661-d78154c9a05d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.901145 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/17dd4503-6f0a-48f1-b661-d78154c9a05d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.901165 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-779g4\" (UniqueName: \"kubernetes.io/projected/17dd4503-6f0a-48f1-b661-d78154c9a05d-kube-api-access-779g4\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:21 crc kubenswrapper[4739]: I1201 15:51:21.901188 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17dd4503-6f0a-48f1-b661-d78154c9a05d-config\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.002576 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-496zm\" (UniqueName: \"kubernetes.io/projected/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-kube-api-access-496zm\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.002623 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-779g4\" (UniqueName: \"kubernetes.io/projected/17dd4503-6f0a-48f1-b661-d78154c9a05d-kube-api-access-779g4\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.002648 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.002753 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.002787 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17dd4503-6f0a-48f1-b661-d78154c9a05d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.002813 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/17dd4503-6f0a-48f1-b661-d78154c9a05d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.002833 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.002875 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-config\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.002895 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/17dd4503-6f0a-48f1-b661-d78154c9a05d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.002965 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.002982 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.003125 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/17dd4503-6f0a-48f1-b661-d78154c9a05d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.003130 4739 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.003478 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/17dd4503-6f0a-48f1-b661-d78154c9a05d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.003144 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17dd4503-6f0a-48f1-b661-d78154c9a05d-config\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.003562 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.003617 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.003672 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17dd4503-6f0a-48f1-b661-d78154c9a05d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.003825 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17dd4503-6f0a-48f1-b661-d78154c9a05d-config\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.004358 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17dd4503-6f0a-48f1-b661-d78154c9a05d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.014241 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/17dd4503-6f0a-48f1-b661-d78154c9a05d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.018096 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/17dd4503-6f0a-48f1-b661-d78154c9a05d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.018339 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17dd4503-6f0a-48f1-b661-d78154c9a05d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.021514 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-779g4\" (UniqueName: \"kubernetes.io/projected/17dd4503-6f0a-48f1-b661-d78154c9a05d-kube-api-access-779g4\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.027281 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"17dd4503-6f0a-48f1-b661-d78154c9a05d\") " pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.054992 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.105401 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.105453 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.105491 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.105517 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.105568 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-496zm\" (UniqueName: \"kubernetes.io/projected/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-kube-api-access-496zm\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.105586 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.105623 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.105639 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-config\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.106448 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-config\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.106542 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.106703 4739 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.107316 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.109110 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.110697 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.111711 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.130548 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.132947 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-496zm\" (UniqueName: \"kubernetes.io/projected/723e3e8b-e65e-46d6-9cd5-11b0d176e8c7-kube-api-access-496zm\") pod \"ovsdbserver-sb-0\" (UID: \"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7\") " pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:22 crc kubenswrapper[4739]: I1201 15:51:22.193500 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:23 crc kubenswrapper[4739]: E1201 15:51:23.793130 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 15:51:23 crc kubenswrapper[4739]: E1201 15:51:23.793886 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b7kxb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-h4ntd_openstack(09152c95-880e-45e7-8ba8-d247d5c5cfc3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 15:51:23 crc kubenswrapper[4739]: E1201 15:51:23.795234 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" podUID="09152c95-880e-45e7-8ba8-d247d5c5cfc3" Dec 01 15:51:23 crc kubenswrapper[4739]: E1201 15:51:23.870563 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 15:51:23 crc kubenswrapper[4739]: E1201 15:51:23.870929 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4m57q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-499zj_openstack(68e76809-d9d8-4d93-8213-cde5d7eb5852): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 15:51:23 crc kubenswrapper[4739]: E1201 15:51:23.875487 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-499zj" podUID="68e76809-d9d8-4d93-8213-cde5d7eb5852" Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.202804 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.279511 4739 generic.go:334] "Generic (PLEG): container finished" podID="6470b09d-b153-4b28-96d4-9425fbbf9e18" containerID="9bae3574cabd9bbc9d32a3ae88a3f18a2fc56f6f43c1caba02f99e7f6574076c" exitCode=0 Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.279589 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" event={"ID":"6470b09d-b153-4b28-96d4-9425fbbf9e18","Type":"ContainerDied","Data":"9bae3574cabd9bbc9d32a3ae88a3f18a2fc56f6f43c1caba02f99e7f6574076c"} Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.283300 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f128863-acb7-4080-829a-01bd78766374","Type":"ContainerStarted","Data":"38694cde5eeb8be42ab1b55e61dfbe17b18565af43593c8d0a588ef65ee37e97"} Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.285064 4739 generic.go:334] "Generic (PLEG): container finished" podID="5ccf3a0a-3522-4825-9a2b-b926a404e642" containerID="5914e17436fcca015a5eb42feceb74fd8211554151454fdb500b0f976184d0a5" exitCode=0 Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.285130 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" event={"ID":"5ccf3a0a-3522-4825-9a2b-b926a404e642","Type":"ContainerDied","Data":"5914e17436fcca015a5eb42feceb74fd8211554151454fdb500b0f976184d0a5"} Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.369302 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 15:51:24 crc kubenswrapper[4739]: W1201 15:51:24.381178 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7301c45d_f6e6_4d47_b75d_d381033edb1d.slice/crio-240e385fbf080f0c4275d741e1f7f43c093394b6faec820faa5b900f58dfa573 WatchSource:0}: Error finding container 240e385fbf080f0c4275d741e1f7f43c093394b6faec820faa5b900f58dfa573: Status 404 returned error can't find the container with id 240e385fbf080f0c4275d741e1f7f43c093394b6faec820faa5b900f58dfa573 Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.738234 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-499zj" Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.746920 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.767301 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.772668 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68e76809-d9d8-4d93-8213-cde5d7eb5852-config\") pod \"68e76809-d9d8-4d93-8213-cde5d7eb5852\" (UID: \"68e76809-d9d8-4d93-8213-cde5d7eb5852\") " Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.772741 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09152c95-880e-45e7-8ba8-d247d5c5cfc3-config\") pod \"09152c95-880e-45e7-8ba8-d247d5c5cfc3\" (UID: \"09152c95-880e-45e7-8ba8-d247d5c5cfc3\") " Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.772784 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7kxb\" (UniqueName: \"kubernetes.io/projected/09152c95-880e-45e7-8ba8-d247d5c5cfc3-kube-api-access-b7kxb\") pod \"09152c95-880e-45e7-8ba8-d247d5c5cfc3\" (UID: \"09152c95-880e-45e7-8ba8-d247d5c5cfc3\") " Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.772827 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09152c95-880e-45e7-8ba8-d247d5c5cfc3-dns-svc\") pod \"09152c95-880e-45e7-8ba8-d247d5c5cfc3\" (UID: \"09152c95-880e-45e7-8ba8-d247d5c5cfc3\") " Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.772855 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4m57q\" (UniqueName: \"kubernetes.io/projected/68e76809-d9d8-4d93-8213-cde5d7eb5852-kube-api-access-4m57q\") pod \"68e76809-d9d8-4d93-8213-cde5d7eb5852\" (UID: \"68e76809-d9d8-4d93-8213-cde5d7eb5852\") " Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.773210 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09152c95-880e-45e7-8ba8-d247d5c5cfc3-config" (OuterVolumeSpecName: "config") pod "09152c95-880e-45e7-8ba8-d247d5c5cfc3" (UID: "09152c95-880e-45e7-8ba8-d247d5c5cfc3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.773338 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68e76809-d9d8-4d93-8213-cde5d7eb5852-config" (OuterVolumeSpecName: "config") pod "68e76809-d9d8-4d93-8213-cde5d7eb5852" (UID: "68e76809-d9d8-4d93-8213-cde5d7eb5852"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.773628 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09152c95-880e-45e7-8ba8-d247d5c5cfc3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "09152c95-880e-45e7-8ba8-d247d5c5cfc3" (UID: "09152c95-880e-45e7-8ba8-d247d5c5cfc3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.779005 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68e76809-d9d8-4d93-8213-cde5d7eb5852-kube-api-access-4m57q" (OuterVolumeSpecName: "kube-api-access-4m57q") pod "68e76809-d9d8-4d93-8213-cde5d7eb5852" (UID: "68e76809-d9d8-4d93-8213-cde5d7eb5852"). InnerVolumeSpecName "kube-api-access-4m57q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.779779 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09152c95-880e-45e7-8ba8-d247d5c5cfc3-kube-api-access-b7kxb" (OuterVolumeSpecName: "kube-api-access-b7kxb") pod "09152c95-880e-45e7-8ba8-d247d5c5cfc3" (UID: "09152c95-880e-45e7-8ba8-d247d5c5cfc3"). InnerVolumeSpecName "kube-api-access-b7kxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:51:24 crc kubenswrapper[4739]: W1201 15:51:24.814570 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51f0a0d6_a82c_495d_9cd4_141c824e6453.slice/crio-b34193f93ed8c47508313827e35ccc38d972ecb418051376c9b13caa2420cf72 WatchSource:0}: Error finding container b34193f93ed8c47508313827e35ccc38d972ecb418051376c9b13caa2420cf72: Status 404 returned error can't find the container with id b34193f93ed8c47508313827e35ccc38d972ecb418051376c9b13caa2420cf72 Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.818065 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ltrp"] Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.831250 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 15:51:24 crc kubenswrapper[4739]: W1201 15:51:24.833943 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8227dcb8_9b5f_4348_bba7_1560fa84861a.slice/crio-fa9141677f6243abe5f8d3009830d40faa7a4384dd0ba90615b0b21414dc2ebf WatchSource:0}: Error finding container fa9141677f6243abe5f8d3009830d40faa7a4384dd0ba90615b0b21414dc2ebf: Status 404 returned error can't find the container with id fa9141677f6243abe5f8d3009830d40faa7a4384dd0ba90615b0b21414dc2ebf Dec 01 15:51:24 crc kubenswrapper[4739]: W1201 15:51:24.836294 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5323c11_dc67_47e7_83be_4b8a52f4ebef.slice/crio-9e0217dba5a915160d87b71258584cc2df829bdce55763b725844652ad326899 WatchSource:0}: Error finding container 9e0217dba5a915160d87b71258584cc2df829bdce55763b725844652ad326899: Status 404 returned error can't find the container with id 9e0217dba5a915160d87b71258584cc2df829bdce55763b725844652ad326899 Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.842587 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.851044 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.875425 4739 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09152c95-880e-45e7-8ba8-d247d5c5cfc3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.875465 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4m57q\" (UniqueName: \"kubernetes.io/projected/68e76809-d9d8-4d93-8213-cde5d7eb5852-kube-api-access-4m57q\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.875480 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68e76809-d9d8-4d93-8213-cde5d7eb5852-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.875557 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09152c95-880e-45e7-8ba8-d247d5c5cfc3-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.875575 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7kxb\" (UniqueName: \"kubernetes.io/projected/09152c95-880e-45e7-8ba8-d247d5c5cfc3-kube-api-access-b7kxb\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:24 crc kubenswrapper[4739]: I1201 15:51:24.954086 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-pr8vl"] Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.293921 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" event={"ID":"09152c95-880e-45e7-8ba8-d247d5c5cfc3","Type":"ContainerDied","Data":"444e239232841adb91349c692d911fb4c78b4fac360dab2df2240da16a43c8ca"} Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.294119 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-h4ntd" Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.304260 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"7301c45d-f6e6-4d47-b75d-d381033edb1d","Type":"ContainerStarted","Data":"240e385fbf080f0c4275d741e1f7f43c093394b6faec820faa5b900f58dfa573"} Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.306157 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pr8vl" event={"ID":"0bed2a41-e961-4361-9316-2f29b6c0b0b5","Type":"ContainerStarted","Data":"c8e5b737410bf9b0b96d2c04018b8295946cbe828556f972612ad26afba97f63"} Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.308352 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" event={"ID":"5ccf3a0a-3522-4825-9a2b-b926a404e642","Type":"ContainerStarted","Data":"a2ba079aadc46ec185779d97ba97fc2b0f16b8fe0098092fc6835b6085a55036"} Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.308533 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.309525 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a5323c11-dc67-47e7-83be-4b8a52f4ebef","Type":"ContainerStarted","Data":"9e0217dba5a915160d87b71258584cc2df829bdce55763b725844652ad326899"} Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.312567 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" event={"ID":"6470b09d-b153-4b28-96d4-9425fbbf9e18","Type":"ContainerStarted","Data":"1687071d6eb0bddfabf56b5d69c125b45c6bc888743d57196823b4a6b765b851"} Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.312657 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.314343 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-499zj" event={"ID":"68e76809-d9d8-4d93-8213-cde5d7eb5852","Type":"ContainerDied","Data":"1424f44cb82bb0afe6a5c115db913a495b78ab18e897773b3c295feb14fc2f12"} Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.314382 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-499zj" Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.315409 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8227dcb8-9b5f-4348-bba7-1560fa84861a","Type":"ContainerStarted","Data":"fa9141677f6243abe5f8d3009830d40faa7a4384dd0ba90615b0b21414dc2ebf"} Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.322355 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ltrp" event={"ID":"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15","Type":"ContainerStarted","Data":"0661ad07c9696173c2c71e6992b03d5ebdb6ffe99465e8a3a126901f24992eb9"} Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.323720 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"51f0a0d6-a82c-495d-9cd4-141c824e6453","Type":"ContainerStarted","Data":"b34193f93ed8c47508313827e35ccc38d972ecb418051376c9b13caa2420cf72"} Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.325190 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a2606100-66af-4acd-98a5-2edcfb281544","Type":"ContainerStarted","Data":"8fcc7a9df5ee3f0cac5d68f0dcac1353400ab87cf57d8f6a7d69317a9d253f9a"} Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.335033 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" podStartSLOduration=6.645844588 podStartE2EDuration="17.335018552s" podCreationTimestamp="2025-12-01 15:51:08 +0000 UTC" firstStartedPulling="2025-12-01 15:51:13.350016157 +0000 UTC m=+975.175762251" lastFinishedPulling="2025-12-01 15:51:24.039190121 +0000 UTC m=+985.864936215" observedRunningTime="2025-12-01 15:51:25.3316928 +0000 UTC m=+987.157438894" watchObservedRunningTime="2025-12-01 15:51:25.335018552 +0000 UTC m=+987.160764656" Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.374686 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" podStartSLOduration=6.681221803 podStartE2EDuration="17.374654659s" podCreationTimestamp="2025-12-01 15:51:08 +0000 UTC" firstStartedPulling="2025-12-01 15:51:13.351659687 +0000 UTC m=+975.177405781" lastFinishedPulling="2025-12-01 15:51:24.045092543 +0000 UTC m=+985.870838637" observedRunningTime="2025-12-01 15:51:25.360577266 +0000 UTC m=+987.186323370" watchObservedRunningTime="2025-12-01 15:51:25.374654659 +0000 UTC m=+987.200400763" Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.426455 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h4ntd"] Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.431588 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h4ntd"] Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.473571 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-499zj"] Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.480171 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-499zj"] Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.647303 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 15:51:25 crc kubenswrapper[4739]: W1201 15:51:25.663543 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17dd4503_6f0a_48f1_b661_d78154c9a05d.slice/crio-bc51145779d3ac05a1a5a43cf5f7f6e85b414f16b1fc7e6a0b2ef9003d6b003d WatchSource:0}: Error finding container bc51145779d3ac05a1a5a43cf5f7f6e85b414f16b1fc7e6a0b2ef9003d6b003d: Status 404 returned error can't find the container with id bc51145779d3ac05a1a5a43cf5f7f6e85b414f16b1fc7e6a0b2ef9003d6b003d Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.940837 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-kznqt"] Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.946193 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.954550 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 01 15:51:25 crc kubenswrapper[4739]: I1201 15:51:25.969907 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kznqt"] Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.007154 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1ea74040-7f82-45a2-a76d-39aca5a88282-ovs-rundir\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.007368 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea74040-7f82-45a2-a76d-39aca5a88282-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.007434 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vkdx\" (UniqueName: \"kubernetes.io/projected/1ea74040-7f82-45a2-a76d-39aca5a88282-kube-api-access-4vkdx\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.007511 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1ea74040-7f82-45a2-a76d-39aca5a88282-ovn-rundir\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.007650 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea74040-7f82-45a2-a76d-39aca5a88282-config\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.007702 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea74040-7f82-45a2-a76d-39aca5a88282-combined-ca-bundle\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.067254 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.103043 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jr7xb"] Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.109697 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1ea74040-7f82-45a2-a76d-39aca5a88282-ovn-rundir\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.109788 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea74040-7f82-45a2-a76d-39aca5a88282-config\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.109821 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea74040-7f82-45a2-a76d-39aca5a88282-combined-ca-bundle\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.109876 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1ea74040-7f82-45a2-a76d-39aca5a88282-ovs-rundir\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.110191 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea74040-7f82-45a2-a76d-39aca5a88282-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.110222 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vkdx\" (UniqueName: \"kubernetes.io/projected/1ea74040-7f82-45a2-a76d-39aca5a88282-kube-api-access-4vkdx\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.110947 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1ea74040-7f82-45a2-a76d-39aca5a88282-ovn-rundir\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.111789 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea74040-7f82-45a2-a76d-39aca5a88282-config\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.114346 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1ea74040-7f82-45a2-a76d-39aca5a88282-ovs-rundir\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.126053 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-sczfq"] Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.129836 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ea74040-7f82-45a2-a76d-39aca5a88282-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.133150 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.146996 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea74040-7f82-45a2-a76d-39aca5a88282-combined-ca-bundle\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.147209 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.152764 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vkdx\" (UniqueName: \"kubernetes.io/projected/1ea74040-7f82-45a2-a76d-39aca5a88282-kube-api-access-4vkdx\") pod \"ovn-controller-metrics-kznqt\" (UID: \"1ea74040-7f82-45a2-a76d-39aca5a88282\") " pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.170909 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-sczfq"] Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.212958 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-sczfq\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.213051 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-sczfq\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.213084 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-config\") pod \"dnsmasq-dns-7fd796d7df-sczfq\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.213156 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z4zz\" (UniqueName: \"kubernetes.io/projected/52c3259e-5284-4903-9cf5-423aae7056a3-kube-api-access-9z4zz\") pod \"dnsmasq-dns-7fd796d7df-sczfq\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.292577 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kznqt" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.314158 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-sczfq\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.314210 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-sczfq\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.314236 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-config\") pod \"dnsmasq-dns-7fd796d7df-sczfq\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.314275 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z4zz\" (UniqueName: \"kubernetes.io/projected/52c3259e-5284-4903-9cf5-423aae7056a3-kube-api-access-9z4zz\") pod \"dnsmasq-dns-7fd796d7df-sczfq\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.315487 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-sczfq\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.316050 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-sczfq\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.316577 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-config\") pod \"dnsmasq-dns-7fd796d7df-sczfq\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.343490 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z4zz\" (UniqueName: \"kubernetes.io/projected/52c3259e-5284-4903-9cf5-423aae7056a3-kube-api-access-9z4zz\") pod \"dnsmasq-dns-7fd796d7df-sczfq\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.345094 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"17dd4503-6f0a-48f1-b661-d78154c9a05d","Type":"ContainerStarted","Data":"bc51145779d3ac05a1a5a43cf5f7f6e85b414f16b1fc7e6a0b2ef9003d6b003d"} Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.351747 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7hv4f"] Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.364347 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kcnlh"] Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.365512 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.370635 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.379989 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kcnlh"] Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.415320 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-kcnlh\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.415620 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-config\") pod \"dnsmasq-dns-86db49b7ff-kcnlh\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.415760 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-kcnlh\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.415848 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-kcnlh\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.416043 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvlr7\" (UniqueName: \"kubernetes.io/projected/2bace9b3-1f79-4e97-8385-580bb4b01ed7-kube-api-access-dvlr7\") pod \"dnsmasq-dns-86db49b7ff-kcnlh\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.489671 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09152c95-880e-45e7-8ba8-d247d5c5cfc3" path="/var/lib/kubelet/pods/09152c95-880e-45e7-8ba8-d247d5c5cfc3/volumes" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.490106 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68e76809-d9d8-4d93-8213-cde5d7eb5852" path="/var/lib/kubelet/pods/68e76809-d9d8-4d93-8213-cde5d7eb5852/volumes" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.498887 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.517306 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvlr7\" (UniqueName: \"kubernetes.io/projected/2bace9b3-1f79-4e97-8385-580bb4b01ed7-kube-api-access-dvlr7\") pod \"dnsmasq-dns-86db49b7ff-kcnlh\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.517474 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-kcnlh\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.517501 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-config\") pod \"dnsmasq-dns-86db49b7ff-kcnlh\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.517536 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-kcnlh\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.517555 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-kcnlh\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.518447 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-kcnlh\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.519190 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-kcnlh\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.519535 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-kcnlh\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.523184 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-config\") pod \"dnsmasq-dns-86db49b7ff-kcnlh\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.535353 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvlr7\" (UniqueName: \"kubernetes.io/projected/2bace9b3-1f79-4e97-8385-580bb4b01ed7-kube-api-access-dvlr7\") pod \"dnsmasq-dns-86db49b7ff-kcnlh\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:26 crc kubenswrapper[4739]: I1201 15:51:26.700844 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:27 crc kubenswrapper[4739]: I1201 15:51:27.359768 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" podUID="5ccf3a0a-3522-4825-9a2b-b926a404e642" containerName="dnsmasq-dns" containerID="cri-o://a2ba079aadc46ec185779d97ba97fc2b0f16b8fe0098092fc6835b6085a55036" gracePeriod=10 Dec 01 15:51:27 crc kubenswrapper[4739]: I1201 15:51:27.359910 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" podUID="6470b09d-b153-4b28-96d4-9425fbbf9e18" containerName="dnsmasq-dns" containerID="cri-o://1687071d6eb0bddfabf56b5d69c125b45c6bc888743d57196823b4a6b765b851" gracePeriod=10 Dec 01 15:51:29 crc kubenswrapper[4739]: I1201 15:51:29.018466 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" podUID="5ccf3a0a-3522-4825-9a2b-b926a404e642" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.97:5353: connect: connection refused" Dec 01 15:51:29 crc kubenswrapper[4739]: W1201 15:51:29.826261 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod723e3e8b_e65e_46d6_9cd5_11b0d176e8c7.slice/crio-2ac2b1d50f88ba5a762bb18db0bd139bfcef95f55c34851033ad71488972360f WatchSource:0}: Error finding container 2ac2b1d50f88ba5a762bb18db0bd139bfcef95f55c34851033ad71488972360f: Status 404 returned error can't find the container with id 2ac2b1d50f88ba5a762bb18db0bd139bfcef95f55c34851033ad71488972360f Dec 01 15:51:30 crc kubenswrapper[4739]: I1201 15:51:30.386270 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kcnlh"] Dec 01 15:51:30 crc kubenswrapper[4739]: I1201 15:51:30.390321 4739 generic.go:334] "Generic (PLEG): container finished" podID="5ccf3a0a-3522-4825-9a2b-b926a404e642" containerID="a2ba079aadc46ec185779d97ba97fc2b0f16b8fe0098092fc6835b6085a55036" exitCode=0 Dec 01 15:51:30 crc kubenswrapper[4739]: I1201 15:51:30.390378 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" event={"ID":"5ccf3a0a-3522-4825-9a2b-b926a404e642","Type":"ContainerDied","Data":"a2ba079aadc46ec185779d97ba97fc2b0f16b8fe0098092fc6835b6085a55036"} Dec 01 15:51:30 crc kubenswrapper[4739]: I1201 15:51:30.393619 4739 generic.go:334] "Generic (PLEG): container finished" podID="6470b09d-b153-4b28-96d4-9425fbbf9e18" containerID="1687071d6eb0bddfabf56b5d69c125b45c6bc888743d57196823b4a6b765b851" exitCode=0 Dec 01 15:51:30 crc kubenswrapper[4739]: I1201 15:51:30.393655 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" event={"ID":"6470b09d-b153-4b28-96d4-9425fbbf9e18","Type":"ContainerDied","Data":"1687071d6eb0bddfabf56b5d69c125b45c6bc888743d57196823b4a6b765b851"} Dec 01 15:51:30 crc kubenswrapper[4739]: I1201 15:51:30.397837 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7","Type":"ContainerStarted","Data":"2ac2b1d50f88ba5a762bb18db0bd139bfcef95f55c34851033ad71488972360f"} Dec 01 15:51:30 crc kubenswrapper[4739]: I1201 15:51:30.544160 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kznqt"] Dec 01 15:51:30 crc kubenswrapper[4739]: I1201 15:51:30.550809 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-sczfq"] Dec 01 15:51:31 crc kubenswrapper[4739]: W1201 15:51:31.740656 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bace9b3_1f79_4e97_8385_580bb4b01ed7.slice/crio-0566478c1d41d6201c071d55c408286199f3458485d78c37f212ad03fe6f07c8 WatchSource:0}: Error finding container 0566478c1d41d6201c071d55c408286199f3458485d78c37f212ad03fe6f07c8: Status 404 returned error can't find the container with id 0566478c1d41d6201c071d55c408286199f3458485d78c37f212ad03fe6f07c8 Dec 01 15:51:31 crc kubenswrapper[4739]: I1201 15:51:31.787657 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" Dec 01 15:51:31 crc kubenswrapper[4739]: I1201 15:51:31.790854 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" Dec 01 15:51:31 crc kubenswrapper[4739]: I1201 15:51:31.918173 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ccf3a0a-3522-4825-9a2b-b926a404e642-config\") pod \"5ccf3a0a-3522-4825-9a2b-b926a404e642\" (UID: \"5ccf3a0a-3522-4825-9a2b-b926a404e642\") " Dec 01 15:51:31 crc kubenswrapper[4739]: I1201 15:51:31.918744 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbg58\" (UniqueName: \"kubernetes.io/projected/6470b09d-b153-4b28-96d4-9425fbbf9e18-kube-api-access-bbg58\") pod \"6470b09d-b153-4b28-96d4-9425fbbf9e18\" (UID: \"6470b09d-b153-4b28-96d4-9425fbbf9e18\") " Dec 01 15:51:31 crc kubenswrapper[4739]: I1201 15:51:31.918825 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6470b09d-b153-4b28-96d4-9425fbbf9e18-config\") pod \"6470b09d-b153-4b28-96d4-9425fbbf9e18\" (UID: \"6470b09d-b153-4b28-96d4-9425fbbf9e18\") " Dec 01 15:51:31 crc kubenswrapper[4739]: I1201 15:51:31.918882 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvgn6\" (UniqueName: \"kubernetes.io/projected/5ccf3a0a-3522-4825-9a2b-b926a404e642-kube-api-access-cvgn6\") pod \"5ccf3a0a-3522-4825-9a2b-b926a404e642\" (UID: \"5ccf3a0a-3522-4825-9a2b-b926a404e642\") " Dec 01 15:51:31 crc kubenswrapper[4739]: I1201 15:51:31.918909 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ccf3a0a-3522-4825-9a2b-b926a404e642-dns-svc\") pod \"5ccf3a0a-3522-4825-9a2b-b926a404e642\" (UID: \"5ccf3a0a-3522-4825-9a2b-b926a404e642\") " Dec 01 15:51:31 crc kubenswrapper[4739]: I1201 15:51:31.918956 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6470b09d-b153-4b28-96d4-9425fbbf9e18-dns-svc\") pod \"6470b09d-b153-4b28-96d4-9425fbbf9e18\" (UID: \"6470b09d-b153-4b28-96d4-9425fbbf9e18\") " Dec 01 15:51:31 crc kubenswrapper[4739]: I1201 15:51:31.925690 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ccf3a0a-3522-4825-9a2b-b926a404e642-kube-api-access-cvgn6" (OuterVolumeSpecName: "kube-api-access-cvgn6") pod "5ccf3a0a-3522-4825-9a2b-b926a404e642" (UID: "5ccf3a0a-3522-4825-9a2b-b926a404e642"). InnerVolumeSpecName "kube-api-access-cvgn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:51:31 crc kubenswrapper[4739]: I1201 15:51:31.926010 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6470b09d-b153-4b28-96d4-9425fbbf9e18-kube-api-access-bbg58" (OuterVolumeSpecName: "kube-api-access-bbg58") pod "6470b09d-b153-4b28-96d4-9425fbbf9e18" (UID: "6470b09d-b153-4b28-96d4-9425fbbf9e18"). InnerVolumeSpecName "kube-api-access-bbg58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:51:31 crc kubenswrapper[4739]: I1201 15:51:31.961321 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6470b09d-b153-4b28-96d4-9425fbbf9e18-config" (OuterVolumeSpecName: "config") pod "6470b09d-b153-4b28-96d4-9425fbbf9e18" (UID: "6470b09d-b153-4b28-96d4-9425fbbf9e18"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:31 crc kubenswrapper[4739]: I1201 15:51:31.965647 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6470b09d-b153-4b28-96d4-9425fbbf9e18-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6470b09d-b153-4b28-96d4-9425fbbf9e18" (UID: "6470b09d-b153-4b28-96d4-9425fbbf9e18"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:31 crc kubenswrapper[4739]: I1201 15:51:31.969601 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ccf3a0a-3522-4825-9a2b-b926a404e642-config" (OuterVolumeSpecName: "config") pod "5ccf3a0a-3522-4825-9a2b-b926a404e642" (UID: "5ccf3a0a-3522-4825-9a2b-b926a404e642"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:31 crc kubenswrapper[4739]: I1201 15:51:31.970847 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ccf3a0a-3522-4825-9a2b-b926a404e642-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5ccf3a0a-3522-4825-9a2b-b926a404e642" (UID: "5ccf3a0a-3522-4825-9a2b-b926a404e642"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.020631 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvgn6\" (UniqueName: \"kubernetes.io/projected/5ccf3a0a-3522-4825-9a2b-b926a404e642-kube-api-access-cvgn6\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.020663 4739 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ccf3a0a-3522-4825-9a2b-b926a404e642-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.020673 4739 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6470b09d-b153-4b28-96d4-9425fbbf9e18-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.020684 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ccf3a0a-3522-4825-9a2b-b926a404e642-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.020693 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbg58\" (UniqueName: \"kubernetes.io/projected/6470b09d-b153-4b28-96d4-9425fbbf9e18-kube-api-access-bbg58\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.020703 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6470b09d-b153-4b28-96d4-9425fbbf9e18-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.414837 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" event={"ID":"2bace9b3-1f79-4e97-8385-580bb4b01ed7","Type":"ContainerStarted","Data":"0566478c1d41d6201c071d55c408286199f3458485d78c37f212ad03fe6f07c8"} Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.416700 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" event={"ID":"5ccf3a0a-3522-4825-9a2b-b926a404e642","Type":"ContainerDied","Data":"d187bad60c30604096f9d6b070ef1f7132777bce3ebb289f6b1f96c7b90067ea"} Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.416729 4739 scope.go:117] "RemoveContainer" containerID="a2ba079aadc46ec185779d97ba97fc2b0f16b8fe0098092fc6835b6085a55036" Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.416760 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-jr7xb" Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.417939 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kznqt" event={"ID":"1ea74040-7f82-45a2-a76d-39aca5a88282","Type":"ContainerStarted","Data":"14ce5bcc4d5c19aa3af008438f7de7af3c384daf6739a61e9269d5401cb27949"} Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.420723 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" event={"ID":"6470b09d-b153-4b28-96d4-9425fbbf9e18","Type":"ContainerDied","Data":"d629e6e613ec25d499029a7849871a0ea706363a73d8f9f758e4fb0b9a9c1e1a"} Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.420825 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7hv4f" Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.463652 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jr7xb"] Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.470321 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jr7xb"] Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.476926 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7hv4f"] Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.486393 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ccf3a0a-3522-4825-9a2b-b926a404e642" path="/var/lib/kubelet/pods/5ccf3a0a-3522-4825-9a2b-b926a404e642/volumes" Dec 01 15:51:32 crc kubenswrapper[4739]: I1201 15:51:32.486989 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7hv4f"] Dec 01 15:51:34 crc kubenswrapper[4739]: I1201 15:51:34.488448 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6470b09d-b153-4b28-96d4-9425fbbf9e18" path="/var/lib/kubelet/pods/6470b09d-b153-4b28-96d4-9425fbbf9e18/volumes" Dec 01 15:51:34 crc kubenswrapper[4739]: W1201 15:51:34.944298 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52c3259e_5284_4903_9cf5_423aae7056a3.slice/crio-fffcf2f6743757551c807a00d395888d0f65e88ed475416b04e91d4274def35e WatchSource:0}: Error finding container fffcf2f6743757551c807a00d395888d0f65e88ed475416b04e91d4274def35e: Status 404 returned error can't find the container with id fffcf2f6743757551c807a00d395888d0f65e88ed475416b04e91d4274def35e Dec 01 15:51:35 crc kubenswrapper[4739]: I1201 15:51:35.458045 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" event={"ID":"52c3259e-5284-4903-9cf5-423aae7056a3","Type":"ContainerStarted","Data":"fffcf2f6743757551c807a00d395888d0f65e88ed475416b04e91d4274def35e"} Dec 01 15:51:35 crc kubenswrapper[4739]: I1201 15:51:35.716177 4739 scope.go:117] "RemoveContainer" containerID="5914e17436fcca015a5eb42feceb74fd8211554151454fdb500b0f976184d0a5" Dec 01 15:51:36 crc kubenswrapper[4739]: I1201 15:51:36.542056 4739 scope.go:117] "RemoveContainer" containerID="1687071d6eb0bddfabf56b5d69c125b45c6bc888743d57196823b4a6b765b851" Dec 01 15:51:38 crc kubenswrapper[4739]: I1201 15:51:38.071945 4739 scope.go:117] "RemoveContainer" containerID="9bae3574cabd9bbc9d32a3ae88a3f18a2fc56f6f43c1caba02f99e7f6574076c" Dec 01 15:51:38 crc kubenswrapper[4739]: I1201 15:51:38.520802 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 01 15:51:38 crc kubenswrapper[4739]: I1201 15:51:38.521608 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"7301c45d-f6e6-4d47-b75d-d381033edb1d","Type":"ContainerStarted","Data":"0adf1068bbba1c2249cb60e256ada8f0950ed146441b7698c9ebcee8e69471ae"} Dec 01 15:51:38 crc kubenswrapper[4739]: I1201 15:51:38.584584 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=15.302444684 podStartE2EDuration="26.584563196s" podCreationTimestamp="2025-12-01 15:51:12 +0000 UTC" firstStartedPulling="2025-12-01 15:51:24.384458929 +0000 UTC m=+986.210205023" lastFinishedPulling="2025-12-01 15:51:35.666577441 +0000 UTC m=+997.492323535" observedRunningTime="2025-12-01 15:51:38.574464815 +0000 UTC m=+1000.400210919" watchObservedRunningTime="2025-12-01 15:51:38.584563196 +0000 UTC m=+1000.410309300" Dec 01 15:51:39 crc kubenswrapper[4739]: I1201 15:51:39.524137 4739 generic.go:334] "Generic (PLEG): container finished" podID="52c3259e-5284-4903-9cf5-423aae7056a3" containerID="5ed12322d0d858aa9ada134710e478bb7f37633013b503fb9332c0306a7e5532" exitCode=0 Dec 01 15:51:39 crc kubenswrapper[4739]: I1201 15:51:39.524319 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" event={"ID":"52c3259e-5284-4903-9cf5-423aae7056a3","Type":"ContainerDied","Data":"5ed12322d0d858aa9ada134710e478bb7f37633013b503fb9332c0306a7e5532"} Dec 01 15:51:39 crc kubenswrapper[4739]: I1201 15:51:39.527391 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a5323c11-dc67-47e7-83be-4b8a52f4ebef","Type":"ContainerStarted","Data":"d7ef1f29e981e8849b2f70f5b9dfabdc92a5b387c72e4cfb320eb4e0ae33941d"} Dec 01 15:51:39 crc kubenswrapper[4739]: I1201 15:51:39.532520 4739 generic.go:334] "Generic (PLEG): container finished" podID="2bace9b3-1f79-4e97-8385-580bb4b01ed7" containerID="1b699e8807e1dc3e7d9809f3729237f47119de43962ade0d21fab5808e4b1a6a" exitCode=0 Dec 01 15:51:39 crc kubenswrapper[4739]: I1201 15:51:39.532651 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" event={"ID":"2bace9b3-1f79-4e97-8385-580bb4b01ed7","Type":"ContainerDied","Data":"1b699e8807e1dc3e7d9809f3729237f47119de43962ade0d21fab5808e4b1a6a"} Dec 01 15:51:39 crc kubenswrapper[4739]: I1201 15:51:39.538941 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7","Type":"ContainerStarted","Data":"7a658f85d092c5b808960b439dddab436c14f645d1696cbb55f1b456a3bc69eb"} Dec 01 15:51:39 crc kubenswrapper[4739]: I1201 15:51:39.543620 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"17dd4503-6f0a-48f1-b661-d78154c9a05d","Type":"ContainerStarted","Data":"6567f50dbd57a954f82b071a9813879850766e45a70ea2b29566e117c993b661"} Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.585596 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ltrp" event={"ID":"10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15","Type":"ContainerStarted","Data":"2e5eb22a6a95469e17cce7c561bf614afb2d6741935f19dd2dcccf22020f4d22"} Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.586856 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-4ltrp" Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.589684 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"51f0a0d6-a82c-495d-9cd4-141c824e6453","Type":"ContainerStarted","Data":"370345a48750c51d8deaa0f4eebdbea37cf4b8ea971f93c0a9e129e6f3f76995"} Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.590390 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.594399 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a2606100-66af-4acd-98a5-2edcfb281544","Type":"ContainerStarted","Data":"184f65dcda1a366c8645ddf60dcc32681e541fb0943f2bcdab6db7840d07555c"} Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.598348 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" event={"ID":"52c3259e-5284-4903-9cf5-423aae7056a3","Type":"ContainerStarted","Data":"d9403ab078eb8ffcac8baab7a7b8291bb41020364e27ca67011c2e49fef635cf"} Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.598472 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.603864 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kznqt" event={"ID":"1ea74040-7f82-45a2-a76d-39aca5a88282","Type":"ContainerStarted","Data":"6640cb2cf55f98914c8976bf546b9bb53058ac3e9eb651080da1a576e2e566d7"} Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.604323 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-4ltrp" podStartSLOduration=9.937130904 podStartE2EDuration="21.604298135s" podCreationTimestamp="2025-12-01 15:51:19 +0000 UTC" firstStartedPulling="2025-12-01 15:51:24.830690428 +0000 UTC m=+986.656436522" lastFinishedPulling="2025-12-01 15:51:36.497857619 +0000 UTC m=+998.323603753" observedRunningTime="2025-12-01 15:51:40.603812681 +0000 UTC m=+1002.429558775" watchObservedRunningTime="2025-12-01 15:51:40.604298135 +0000 UTC m=+1002.430044239" Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.606537 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" event={"ID":"2bace9b3-1f79-4e97-8385-580bb4b01ed7","Type":"ContainerStarted","Data":"c31c7d1d2d67a9db18e15f8088cec205d2b7d85e64134a4ac79d468dd1d0e4ec"} Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.606666 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.608712 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f128863-acb7-4080-829a-01bd78766374","Type":"ContainerStarted","Data":"17bbc7387e706b89896707fd45f1fbe1e75abe86f809cbdf894a0b9ec75c4520"} Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.613980 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"17dd4503-6f0a-48f1-b661-d78154c9a05d","Type":"ContainerStarted","Data":"6445115e057a3c9d701d80b4a5e37cf0fbff91ab377a883784d44785c335aea2"} Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.617064 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=12.231464921 podStartE2EDuration="26.616647605s" podCreationTimestamp="2025-12-01 15:51:14 +0000 UTC" firstStartedPulling="2025-12-01 15:51:24.823752304 +0000 UTC m=+986.649498398" lastFinishedPulling="2025-12-01 15:51:39.208934988 +0000 UTC m=+1001.034681082" observedRunningTime="2025-12-01 15:51:40.615125978 +0000 UTC m=+1002.440872072" watchObservedRunningTime="2025-12-01 15:51:40.616647605 +0000 UTC m=+1002.442393699" Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.623015 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"723e3e8b-e65e-46d6-9cd5-11b0d176e8c7","Type":"ContainerStarted","Data":"2c8260028c8c963d170d844ee6d40d81eb539cfad697182014ebb9d6ba50a784"} Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.624304 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pr8vl" event={"ID":"0bed2a41-e961-4361-9316-2f29b6c0b0b5","Type":"ContainerStarted","Data":"b8b88391c304b4deb726d49ed0cec80b080df66606e594d5d68e078848cabedf"} Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.658743 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" podStartSLOduration=14.658727887 podStartE2EDuration="14.658727887s" podCreationTimestamp="2025-12-01 15:51:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:51:40.657067117 +0000 UTC m=+1002.482813211" watchObservedRunningTime="2025-12-01 15:51:40.658727887 +0000 UTC m=+1002.484473981" Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.757666 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=8.689699973 podStartE2EDuration="20.757648617s" podCreationTimestamp="2025-12-01 15:51:20 +0000 UTC" firstStartedPulling="2025-12-01 15:51:25.670634252 +0000 UTC m=+987.496380346" lastFinishedPulling="2025-12-01 15:51:37.738582896 +0000 UTC m=+999.564328990" observedRunningTime="2025-12-01 15:51:40.751520038 +0000 UTC m=+1002.577266142" watchObservedRunningTime="2025-12-01 15:51:40.757648617 +0000 UTC m=+1002.583394701" Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.773774 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-kznqt" podStartSLOduration=8.453533694 podStartE2EDuration="15.773756881s" podCreationTimestamp="2025-12-01 15:51:25 +0000 UTC" firstStartedPulling="2025-12-01 15:51:31.725166936 +0000 UTC m=+993.550913030" lastFinishedPulling="2025-12-01 15:51:39.045390113 +0000 UTC m=+1000.871136217" observedRunningTime="2025-12-01 15:51:40.766809077 +0000 UTC m=+1002.592555171" watchObservedRunningTime="2025-12-01 15:51:40.773756881 +0000 UTC m=+1002.599502975" Dec 01 15:51:40 crc kubenswrapper[4739]: I1201 15:51:40.817750 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" podStartSLOduration=14.817727382 podStartE2EDuration="14.817727382s" podCreationTimestamp="2025-12-01 15:51:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:51:40.784075938 +0000 UTC m=+1002.609822032" watchObservedRunningTime="2025-12-01 15:51:40.817727382 +0000 UTC m=+1002.643473476" Dec 01 15:51:41 crc kubenswrapper[4739]: I1201 15:51:41.637628 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8227dcb8-9b5f-4348-bba7-1560fa84861a","Type":"ContainerStarted","Data":"81b7dbc26dbad80f9b0b48a2cf196f714e38137e5d371c4ade49d65cb0fd4f41"} Dec 01 15:51:41 crc kubenswrapper[4739]: I1201 15:51:41.640221 4739 generic.go:334] "Generic (PLEG): container finished" podID="0bed2a41-e961-4361-9316-2f29b6c0b0b5" containerID="b8b88391c304b4deb726d49ed0cec80b080df66606e594d5d68e078848cabedf" exitCode=0 Dec 01 15:51:41 crc kubenswrapper[4739]: I1201 15:51:41.640320 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pr8vl" event={"ID":"0bed2a41-e961-4361-9316-2f29b6c0b0b5","Type":"ContainerDied","Data":"b8b88391c304b4deb726d49ed0cec80b080df66606e594d5d68e078848cabedf"} Dec 01 15:51:41 crc kubenswrapper[4739]: I1201 15:51:41.681558 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=14.87329575 podStartE2EDuration="21.68153462s" podCreationTimestamp="2025-12-01 15:51:20 +0000 UTC" firstStartedPulling="2025-12-01 15:51:29.834225653 +0000 UTC m=+991.659971747" lastFinishedPulling="2025-12-01 15:51:36.642464483 +0000 UTC m=+998.468210617" observedRunningTime="2025-12-01 15:51:40.818045612 +0000 UTC m=+1002.643791716" watchObservedRunningTime="2025-12-01 15:51:41.68153462 +0000 UTC m=+1003.507280754" Dec 01 15:51:42 crc kubenswrapper[4739]: I1201 15:51:42.055185 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:42 crc kubenswrapper[4739]: I1201 15:51:42.193972 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:42 crc kubenswrapper[4739]: I1201 15:51:42.650854 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pr8vl" event={"ID":"0bed2a41-e961-4361-9316-2f29b6c0b0b5","Type":"ContainerStarted","Data":"518f27183f7290259ee7712947d0a37d2cea0b2f6a79e7d9bf20d4071c2d6a9a"} Dec 01 15:51:42 crc kubenswrapper[4739]: I1201 15:51:42.651118 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pr8vl" event={"ID":"0bed2a41-e961-4361-9316-2f29b6c0b0b5","Type":"ContainerStarted","Data":"d163da930d657904fa1d6620bb6b35c2fa17a26676384abcba550396bcdf2d20"} Dec 01 15:51:42 crc kubenswrapper[4739]: I1201 15:51:42.651135 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:42 crc kubenswrapper[4739]: I1201 15:51:42.651149 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:51:42 crc kubenswrapper[4739]: I1201 15:51:42.652941 4739 generic.go:334] "Generic (PLEG): container finished" podID="a5323c11-dc67-47e7-83be-4b8a52f4ebef" containerID="d7ef1f29e981e8849b2f70f5b9dfabdc92a5b387c72e4cfb320eb4e0ae33941d" exitCode=0 Dec 01 15:51:42 crc kubenswrapper[4739]: I1201 15:51:42.653476 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a5323c11-dc67-47e7-83be-4b8a52f4ebef","Type":"ContainerDied","Data":"d7ef1f29e981e8849b2f70f5b9dfabdc92a5b387c72e4cfb320eb4e0ae33941d"} Dec 01 15:51:42 crc kubenswrapper[4739]: I1201 15:51:42.678831 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-pr8vl" podStartSLOduration=12.598008409 podStartE2EDuration="23.678808917s" podCreationTimestamp="2025-12-01 15:51:19 +0000 UTC" firstStartedPulling="2025-12-01 15:51:24.964114857 +0000 UTC m=+986.789860971" lastFinishedPulling="2025-12-01 15:51:36.044915375 +0000 UTC m=+997.870661479" observedRunningTime="2025-12-01 15:51:42.67207212 +0000 UTC m=+1004.497818234" watchObservedRunningTime="2025-12-01 15:51:42.678808917 +0000 UTC m=+1004.504555011" Dec 01 15:51:43 crc kubenswrapper[4739]: I1201 15:51:43.056236 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:43 crc kubenswrapper[4739]: I1201 15:51:43.112485 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:43 crc kubenswrapper[4739]: I1201 15:51:43.193941 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:43 crc kubenswrapper[4739]: I1201 15:51:43.242148 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:43 crc kubenswrapper[4739]: I1201 15:51:43.299187 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 01 15:51:43 crc kubenswrapper[4739]: I1201 15:51:43.667011 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a5323c11-dc67-47e7-83be-4b8a52f4ebef","Type":"ContainerStarted","Data":"467629453ad4c5aedc6f8b4641fe57ecd0a6039cd99e5c82191ac31b791c136e"} Dec 01 15:51:43 crc kubenswrapper[4739]: I1201 15:51:43.696999 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=21.500473705 podStartE2EDuration="32.696975457s" podCreationTimestamp="2025-12-01 15:51:11 +0000 UTC" firstStartedPulling="2025-12-01 15:51:24.840686335 +0000 UTC m=+986.666432429" lastFinishedPulling="2025-12-01 15:51:36.037188077 +0000 UTC m=+997.862934181" observedRunningTime="2025-12-01 15:51:43.689131186 +0000 UTC m=+1005.514877280" watchObservedRunningTime="2025-12-01 15:51:43.696975457 +0000 UTC m=+1005.522721571" Dec 01 15:51:44 crc kubenswrapper[4739]: I1201 15:51:44.678162 4739 generic.go:334] "Generic (PLEG): container finished" podID="8227dcb8-9b5f-4348-bba7-1560fa84861a" containerID="81b7dbc26dbad80f9b0b48a2cf196f714e38137e5d371c4ade49d65cb0fd4f41" exitCode=0 Dec 01 15:51:44 crc kubenswrapper[4739]: I1201 15:51:44.678212 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8227dcb8-9b5f-4348-bba7-1560fa84861a","Type":"ContainerDied","Data":"81b7dbc26dbad80f9b0b48a2cf196f714e38137e5d371c4ade49d65cb0fd4f41"} Dec 01 15:51:44 crc kubenswrapper[4739]: I1201 15:51:44.739207 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 01 15:51:44 crc kubenswrapper[4739]: I1201 15:51:44.752766 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.041079 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 01 15:51:45 crc kubenswrapper[4739]: E1201 15:51:45.041852 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6470b09d-b153-4b28-96d4-9425fbbf9e18" containerName="init" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.041877 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="6470b09d-b153-4b28-96d4-9425fbbf9e18" containerName="init" Dec 01 15:51:45 crc kubenswrapper[4739]: E1201 15:51:45.041897 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ccf3a0a-3522-4825-9a2b-b926a404e642" containerName="init" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.041905 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ccf3a0a-3522-4825-9a2b-b926a404e642" containerName="init" Dec 01 15:51:45 crc kubenswrapper[4739]: E1201 15:51:45.041930 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ccf3a0a-3522-4825-9a2b-b926a404e642" containerName="dnsmasq-dns" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.041938 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ccf3a0a-3522-4825-9a2b-b926a404e642" containerName="dnsmasq-dns" Dec 01 15:51:45 crc kubenswrapper[4739]: E1201 15:51:45.041960 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6470b09d-b153-4b28-96d4-9425fbbf9e18" containerName="dnsmasq-dns" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.041967 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="6470b09d-b153-4b28-96d4-9425fbbf9e18" containerName="dnsmasq-dns" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.042147 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="6470b09d-b153-4b28-96d4-9425fbbf9e18" containerName="dnsmasq-dns" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.042168 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ccf3a0a-3522-4825-9a2b-b926a404e642" containerName="dnsmasq-dns" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.043135 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.046776 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-pxxpj" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.046903 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.046971 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.047036 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.052362 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.149268 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.149355 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.149507 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-config\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.149561 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.149581 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.149603 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrs9p\" (UniqueName: \"kubernetes.io/projected/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-kube-api-access-rrs9p\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.149731 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-scripts\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.254527 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.254585 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.254627 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-config\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.254661 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.254676 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.254690 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrs9p\" (UniqueName: \"kubernetes.io/projected/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-kube-api-access-rrs9p\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.254711 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-scripts\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.255574 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-scripts\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.255640 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-config\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.255994 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.260308 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.260485 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.262294 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.274210 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrs9p\" (UniqueName: \"kubernetes.io/projected/cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d-kube-api-access-rrs9p\") pod \"ovn-northd-0\" (UID: \"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d\") " pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.280026 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.372859 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.687930 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8227dcb8-9b5f-4348-bba7-1560fa84861a","Type":"ContainerStarted","Data":"fb66c7eab5b9e9986fd8f9d1596a4dab810832a3fcee99241551d0e3555fb9e0"} Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.723416 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=22.81219505 podStartE2EDuration="35.723401511s" podCreationTimestamp="2025-12-01 15:51:10 +0000 UTC" firstStartedPulling="2025-12-01 15:51:24.83729494 +0000 UTC m=+986.663041034" lastFinishedPulling="2025-12-01 15:51:37.748501401 +0000 UTC m=+999.574247495" observedRunningTime="2025-12-01 15:51:45.720552454 +0000 UTC m=+1007.546298578" watchObservedRunningTime="2025-12-01 15:51:45.723401511 +0000 UTC m=+1007.549147605" Dec 01 15:51:45 crc kubenswrapper[4739]: I1201 15:51:45.843086 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 15:51:46 crc kubenswrapper[4739]: I1201 15:51:46.500757 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:46 crc kubenswrapper[4739]: I1201 15:51:46.700068 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d","Type":"ContainerStarted","Data":"6b6ca7065d825dc83eed28f50c93c131c33a9b5212ed12456e98fe3c91b35463"} Dec 01 15:51:46 crc kubenswrapper[4739]: I1201 15:51:46.703469 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:51:46 crc kubenswrapper[4739]: I1201 15:51:46.771365 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-sczfq"] Dec 01 15:51:46 crc kubenswrapper[4739]: I1201 15:51:46.771582 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" podUID="52c3259e-5284-4903-9cf5-423aae7056a3" containerName="dnsmasq-dns" containerID="cri-o://d9403ab078eb8ffcac8baab7a7b8291bb41020364e27ca67011c2e49fef635cf" gracePeriod=10 Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.436057 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.604186 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-dns-svc\") pod \"52c3259e-5284-4903-9cf5-423aae7056a3\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.604340 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z4zz\" (UniqueName: \"kubernetes.io/projected/52c3259e-5284-4903-9cf5-423aae7056a3-kube-api-access-9z4zz\") pod \"52c3259e-5284-4903-9cf5-423aae7056a3\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.604381 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-ovsdbserver-nb\") pod \"52c3259e-5284-4903-9cf5-423aae7056a3\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.604717 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-config\") pod \"52c3259e-5284-4903-9cf5-423aae7056a3\" (UID: \"52c3259e-5284-4903-9cf5-423aae7056a3\") " Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.607787 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52c3259e-5284-4903-9cf5-423aae7056a3-kube-api-access-9z4zz" (OuterVolumeSpecName: "kube-api-access-9z4zz") pod "52c3259e-5284-4903-9cf5-423aae7056a3" (UID: "52c3259e-5284-4903-9cf5-423aae7056a3"). InnerVolumeSpecName "kube-api-access-9z4zz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.639805 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "52c3259e-5284-4903-9cf5-423aae7056a3" (UID: "52c3259e-5284-4903-9cf5-423aae7056a3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.649969 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "52c3259e-5284-4903-9cf5-423aae7056a3" (UID: "52c3259e-5284-4903-9cf5-423aae7056a3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.652783 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-config" (OuterVolumeSpecName: "config") pod "52c3259e-5284-4903-9cf5-423aae7056a3" (UID: "52c3259e-5284-4903-9cf5-423aae7056a3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.706625 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9z4zz\" (UniqueName: \"kubernetes.io/projected/52c3259e-5284-4903-9cf5-423aae7056a3-kube-api-access-9z4zz\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.706651 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.706661 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.706672 4739 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52c3259e-5284-4903-9cf5-423aae7056a3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.710369 4739 generic.go:334] "Generic (PLEG): container finished" podID="52c3259e-5284-4903-9cf5-423aae7056a3" containerID="d9403ab078eb8ffcac8baab7a7b8291bb41020364e27ca67011c2e49fef635cf" exitCode=0 Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.710443 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.710462 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" event={"ID":"52c3259e-5284-4903-9cf5-423aae7056a3","Type":"ContainerDied","Data":"d9403ab078eb8ffcac8baab7a7b8291bb41020364e27ca67011c2e49fef635cf"} Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.710504 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-sczfq" event={"ID":"52c3259e-5284-4903-9cf5-423aae7056a3","Type":"ContainerDied","Data":"fffcf2f6743757551c807a00d395888d0f65e88ed475416b04e91d4274def35e"} Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.710520 4739 scope.go:117] "RemoveContainer" containerID="d9403ab078eb8ffcac8baab7a7b8291bb41020364e27ca67011c2e49fef635cf" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.714470 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d","Type":"ContainerStarted","Data":"27ed796a9901d8401ca88af748c2a907e973dbb645ea469d23141da61b1e571f"} Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.714504 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d","Type":"ContainerStarted","Data":"6307d5684502c953a8a33487f8026bdfa8155cffc232bb65fb860d8fe190e9a4"} Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.714767 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.734322 4739 scope.go:117] "RemoveContainer" containerID="5ed12322d0d858aa9ada134710e478bb7f37633013b503fb9332c0306a7e5532" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.744962 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.415940067 podStartE2EDuration="2.744940306s" podCreationTimestamp="2025-12-01 15:51:45 +0000 UTC" firstStartedPulling="2025-12-01 15:51:45.851715153 +0000 UTC m=+1007.677461257" lastFinishedPulling="2025-12-01 15:51:47.180715402 +0000 UTC m=+1009.006461496" observedRunningTime="2025-12-01 15:51:47.730501222 +0000 UTC m=+1009.556247336" watchObservedRunningTime="2025-12-01 15:51:47.744940306 +0000 UTC m=+1009.570686410" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.750929 4739 scope.go:117] "RemoveContainer" containerID="d9403ab078eb8ffcac8baab7a7b8291bb41020364e27ca67011c2e49fef635cf" Dec 01 15:51:47 crc kubenswrapper[4739]: E1201 15:51:47.751400 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9403ab078eb8ffcac8baab7a7b8291bb41020364e27ca67011c2e49fef635cf\": container with ID starting with d9403ab078eb8ffcac8baab7a7b8291bb41020364e27ca67011c2e49fef635cf not found: ID does not exist" containerID="d9403ab078eb8ffcac8baab7a7b8291bb41020364e27ca67011c2e49fef635cf" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.751470 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9403ab078eb8ffcac8baab7a7b8291bb41020364e27ca67011c2e49fef635cf"} err="failed to get container status \"d9403ab078eb8ffcac8baab7a7b8291bb41020364e27ca67011c2e49fef635cf\": rpc error: code = NotFound desc = could not find container \"d9403ab078eb8ffcac8baab7a7b8291bb41020364e27ca67011c2e49fef635cf\": container with ID starting with d9403ab078eb8ffcac8baab7a7b8291bb41020364e27ca67011c2e49fef635cf not found: ID does not exist" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.751505 4739 scope.go:117] "RemoveContainer" containerID="5ed12322d0d858aa9ada134710e478bb7f37633013b503fb9332c0306a7e5532" Dec 01 15:51:47 crc kubenswrapper[4739]: E1201 15:51:47.751862 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ed12322d0d858aa9ada134710e478bb7f37633013b503fb9332c0306a7e5532\": container with ID starting with 5ed12322d0d858aa9ada134710e478bb7f37633013b503fb9332c0306a7e5532 not found: ID does not exist" containerID="5ed12322d0d858aa9ada134710e478bb7f37633013b503fb9332c0306a7e5532" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.751929 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ed12322d0d858aa9ada134710e478bb7f37633013b503fb9332c0306a7e5532"} err="failed to get container status \"5ed12322d0d858aa9ada134710e478bb7f37633013b503fb9332c0306a7e5532\": rpc error: code = NotFound desc = could not find container \"5ed12322d0d858aa9ada134710e478bb7f37633013b503fb9332c0306a7e5532\": container with ID starting with 5ed12322d0d858aa9ada134710e478bb7f37633013b503fb9332c0306a7e5532 not found: ID does not exist" Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.758996 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-sczfq"] Dec 01 15:51:47 crc kubenswrapper[4739]: I1201 15:51:47.764611 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-sczfq"] Dec 01 15:51:48 crc kubenswrapper[4739]: I1201 15:51:48.493136 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52c3259e-5284-4903-9cf5-423aae7056a3" path="/var/lib/kubelet/pods/52c3259e-5284-4903-9cf5-423aae7056a3/volumes" Dec 01 15:51:51 crc kubenswrapper[4739]: I1201 15:51:51.514280 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 01 15:51:51 crc kubenswrapper[4739]: I1201 15:51:51.514821 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 01 15:51:51 crc kubenswrapper[4739]: I1201 15:51:51.622518 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 01 15:51:51 crc kubenswrapper[4739]: I1201 15:51:51.850604 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 01 15:51:52 crc kubenswrapper[4739]: I1201 15:51:52.951937 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-ed22-account-create-update-c4lxx"] Dec 01 15:51:52 crc kubenswrapper[4739]: E1201 15:51:52.952312 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52c3259e-5284-4903-9cf5-423aae7056a3" containerName="init" Dec 01 15:51:52 crc kubenswrapper[4739]: I1201 15:51:52.952325 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="52c3259e-5284-4903-9cf5-423aae7056a3" containerName="init" Dec 01 15:51:52 crc kubenswrapper[4739]: E1201 15:51:52.952343 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52c3259e-5284-4903-9cf5-423aae7056a3" containerName="dnsmasq-dns" Dec 01 15:51:52 crc kubenswrapper[4739]: I1201 15:51:52.952350 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="52c3259e-5284-4903-9cf5-423aae7056a3" containerName="dnsmasq-dns" Dec 01 15:51:52 crc kubenswrapper[4739]: I1201 15:51:52.952563 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="52c3259e-5284-4903-9cf5-423aae7056a3" containerName="dnsmasq-dns" Dec 01 15:51:52 crc kubenswrapper[4739]: I1201 15:51:52.953194 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ed22-account-create-update-c4lxx" Dec 01 15:51:52 crc kubenswrapper[4739]: I1201 15:51:52.957764 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 01 15:51:52 crc kubenswrapper[4739]: I1201 15:51:52.971442 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ed22-account-create-update-c4lxx"] Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.015404 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-dxcwt"] Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.016633 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-dxcwt" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.027101 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-dxcwt"] Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.093674 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.093722 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.105324 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcs6h\" (UniqueName: \"kubernetes.io/projected/618420cf-adef-4233-b155-d213621354a0-kube-api-access-kcs6h\") pod \"keystone-ed22-account-create-update-c4lxx\" (UID: \"618420cf-adef-4233-b155-d213621354a0\") " pod="openstack/keystone-ed22-account-create-update-c4lxx" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.105439 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/618420cf-adef-4233-b155-d213621354a0-operator-scripts\") pod \"keystone-ed22-account-create-update-c4lxx\" (UID: \"618420cf-adef-4233-b155-d213621354a0\") " pod="openstack/keystone-ed22-account-create-update-c4lxx" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.105519 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7ff093b-772b-4c09-ba41-6f5fd9324591-operator-scripts\") pod \"keystone-db-create-dxcwt\" (UID: \"f7ff093b-772b-4c09-ba41-6f5fd9324591\") " pod="openstack/keystone-db-create-dxcwt" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.105547 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gptxv\" (UniqueName: \"kubernetes.io/projected/f7ff093b-772b-4c09-ba41-6f5fd9324591-kube-api-access-gptxv\") pod \"keystone-db-create-dxcwt\" (UID: \"f7ff093b-772b-4c09-ba41-6f5fd9324591\") " pod="openstack/keystone-db-create-dxcwt" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.184312 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-skpk2"] Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.185481 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-skpk2" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.194557 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-skpk2"] Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.200313 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.207711 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcs6h\" (UniqueName: \"kubernetes.io/projected/618420cf-adef-4233-b155-d213621354a0-kube-api-access-kcs6h\") pod \"keystone-ed22-account-create-update-c4lxx\" (UID: \"618420cf-adef-4233-b155-d213621354a0\") " pod="openstack/keystone-ed22-account-create-update-c4lxx" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.207865 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/618420cf-adef-4233-b155-d213621354a0-operator-scripts\") pod \"keystone-ed22-account-create-update-c4lxx\" (UID: \"618420cf-adef-4233-b155-d213621354a0\") " pod="openstack/keystone-ed22-account-create-update-c4lxx" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.208060 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7ff093b-772b-4c09-ba41-6f5fd9324591-operator-scripts\") pod \"keystone-db-create-dxcwt\" (UID: \"f7ff093b-772b-4c09-ba41-6f5fd9324591\") " pod="openstack/keystone-db-create-dxcwt" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.208116 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gptxv\" (UniqueName: \"kubernetes.io/projected/f7ff093b-772b-4c09-ba41-6f5fd9324591-kube-api-access-gptxv\") pod \"keystone-db-create-dxcwt\" (UID: \"f7ff093b-772b-4c09-ba41-6f5fd9324591\") " pod="openstack/keystone-db-create-dxcwt" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.209919 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/618420cf-adef-4233-b155-d213621354a0-operator-scripts\") pod \"keystone-ed22-account-create-update-c4lxx\" (UID: \"618420cf-adef-4233-b155-d213621354a0\") " pod="openstack/keystone-ed22-account-create-update-c4lxx" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.210905 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7ff093b-772b-4c09-ba41-6f5fd9324591-operator-scripts\") pod \"keystone-db-create-dxcwt\" (UID: \"f7ff093b-772b-4c09-ba41-6f5fd9324591\") " pod="openstack/keystone-db-create-dxcwt" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.236194 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gptxv\" (UniqueName: \"kubernetes.io/projected/f7ff093b-772b-4c09-ba41-6f5fd9324591-kube-api-access-gptxv\") pod \"keystone-db-create-dxcwt\" (UID: \"f7ff093b-772b-4c09-ba41-6f5fd9324591\") " pod="openstack/keystone-db-create-dxcwt" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.237179 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcs6h\" (UniqueName: \"kubernetes.io/projected/618420cf-adef-4233-b155-d213621354a0-kube-api-access-kcs6h\") pod \"keystone-ed22-account-create-update-c4lxx\" (UID: \"618420cf-adef-4233-b155-d213621354a0\") " pod="openstack/keystone-ed22-account-create-update-c4lxx" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.261370 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-d1ed-account-create-update-64cxx"] Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.264532 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d1ed-account-create-update-64cxx" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.266878 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.273508 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d1ed-account-create-update-64cxx"] Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.281382 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ed22-account-create-update-c4lxx" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.309270 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plvm2\" (UniqueName: \"kubernetes.io/projected/9add3351-b3aa-41a1-8c08-5c10147e2520-kube-api-access-plvm2\") pod \"placement-db-create-skpk2\" (UID: \"9add3351-b3aa-41a1-8c08-5c10147e2520\") " pod="openstack/placement-db-create-skpk2" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.309326 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9add3351-b3aa-41a1-8c08-5c10147e2520-operator-scripts\") pod \"placement-db-create-skpk2\" (UID: \"9add3351-b3aa-41a1-8c08-5c10147e2520\") " pod="openstack/placement-db-create-skpk2" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.334774 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-dxcwt" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.410643 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dceccaf5-f087-477b-b9ac-4a0748136112-operator-scripts\") pod \"placement-d1ed-account-create-update-64cxx\" (UID: \"dceccaf5-f087-477b-b9ac-4a0748136112\") " pod="openstack/placement-d1ed-account-create-update-64cxx" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.411777 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plvm2\" (UniqueName: \"kubernetes.io/projected/9add3351-b3aa-41a1-8c08-5c10147e2520-kube-api-access-plvm2\") pod \"placement-db-create-skpk2\" (UID: \"9add3351-b3aa-41a1-8c08-5c10147e2520\") " pod="openstack/placement-db-create-skpk2" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.411815 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9add3351-b3aa-41a1-8c08-5c10147e2520-operator-scripts\") pod \"placement-db-create-skpk2\" (UID: \"9add3351-b3aa-41a1-8c08-5c10147e2520\") " pod="openstack/placement-db-create-skpk2" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.411913 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxvw5\" (UniqueName: \"kubernetes.io/projected/dceccaf5-f087-477b-b9ac-4a0748136112-kube-api-access-qxvw5\") pod \"placement-d1ed-account-create-update-64cxx\" (UID: \"dceccaf5-f087-477b-b9ac-4a0748136112\") " pod="openstack/placement-d1ed-account-create-update-64cxx" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.412918 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9add3351-b3aa-41a1-8c08-5c10147e2520-operator-scripts\") pod \"placement-db-create-skpk2\" (UID: \"9add3351-b3aa-41a1-8c08-5c10147e2520\") " pod="openstack/placement-db-create-skpk2" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.446059 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plvm2\" (UniqueName: \"kubernetes.io/projected/9add3351-b3aa-41a1-8c08-5c10147e2520-kube-api-access-plvm2\") pod \"placement-db-create-skpk2\" (UID: \"9add3351-b3aa-41a1-8c08-5c10147e2520\") " pod="openstack/placement-db-create-skpk2" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.482928 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-27pv4"] Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.484084 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-27pv4" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.493331 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-27pv4"] Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.506118 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-skpk2" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.514004 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxvw5\" (UniqueName: \"kubernetes.io/projected/dceccaf5-f087-477b-b9ac-4a0748136112-kube-api-access-qxvw5\") pod \"placement-d1ed-account-create-update-64cxx\" (UID: \"dceccaf5-f087-477b-b9ac-4a0748136112\") " pod="openstack/placement-d1ed-account-create-update-64cxx" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.514082 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dceccaf5-f087-477b-b9ac-4a0748136112-operator-scripts\") pod \"placement-d1ed-account-create-update-64cxx\" (UID: \"dceccaf5-f087-477b-b9ac-4a0748136112\") " pod="openstack/placement-d1ed-account-create-update-64cxx" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.514984 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dceccaf5-f087-477b-b9ac-4a0748136112-operator-scripts\") pod \"placement-d1ed-account-create-update-64cxx\" (UID: \"dceccaf5-f087-477b-b9ac-4a0748136112\") " pod="openstack/placement-d1ed-account-create-update-64cxx" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.533315 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxvw5\" (UniqueName: \"kubernetes.io/projected/dceccaf5-f087-477b-b9ac-4a0748136112-kube-api-access-qxvw5\") pod \"placement-d1ed-account-create-update-64cxx\" (UID: \"dceccaf5-f087-477b-b9ac-4a0748136112\") " pod="openstack/placement-d1ed-account-create-update-64cxx" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.604331 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-8170-account-create-update-8jz42"] Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.605331 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8170-account-create-update-8jz42" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.608127 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.613037 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-8170-account-create-update-8jz42"] Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.615047 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bclgq\" (UniqueName: \"kubernetes.io/projected/52f45a77-0faf-42ae-b052-ccb63b092440-kube-api-access-bclgq\") pod \"glance-db-create-27pv4\" (UID: \"52f45a77-0faf-42ae-b052-ccb63b092440\") " pod="openstack/glance-db-create-27pv4" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.615212 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52f45a77-0faf-42ae-b052-ccb63b092440-operator-scripts\") pod \"glance-db-create-27pv4\" (UID: \"52f45a77-0faf-42ae-b052-ccb63b092440\") " pod="openstack/glance-db-create-27pv4" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.714935 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ed22-account-create-update-c4lxx"] Dec 01 15:51:53 crc kubenswrapper[4739]: W1201 15:51:53.716557 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod618420cf_adef_4233_b155_d213621354a0.slice/crio-06447f442fa1852213064c40ed55ae0f7ad4ed60dd89fe0aab732753b6816acf WatchSource:0}: Error finding container 06447f442fa1852213064c40ed55ae0f7ad4ed60dd89fe0aab732753b6816acf: Status 404 returned error can't find the container with id 06447f442fa1852213064c40ed55ae0f7ad4ed60dd89fe0aab732753b6816acf Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.716940 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52f45a77-0faf-42ae-b052-ccb63b092440-operator-scripts\") pod \"glance-db-create-27pv4\" (UID: \"52f45a77-0faf-42ae-b052-ccb63b092440\") " pod="openstack/glance-db-create-27pv4" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.717017 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5e17350-562f-4cd1-a861-0cea798c8528-operator-scripts\") pod \"glance-8170-account-create-update-8jz42\" (UID: \"d5e17350-562f-4cd1-a861-0cea798c8528\") " pod="openstack/glance-8170-account-create-update-8jz42" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.717044 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bclgq\" (UniqueName: \"kubernetes.io/projected/52f45a77-0faf-42ae-b052-ccb63b092440-kube-api-access-bclgq\") pod \"glance-db-create-27pv4\" (UID: \"52f45a77-0faf-42ae-b052-ccb63b092440\") " pod="openstack/glance-db-create-27pv4" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.717074 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6rlx\" (UniqueName: \"kubernetes.io/projected/d5e17350-562f-4cd1-a861-0cea798c8528-kube-api-access-p6rlx\") pod \"glance-8170-account-create-update-8jz42\" (UID: \"d5e17350-562f-4cd1-a861-0cea798c8528\") " pod="openstack/glance-8170-account-create-update-8jz42" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.717820 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52f45a77-0faf-42ae-b052-ccb63b092440-operator-scripts\") pod \"glance-db-create-27pv4\" (UID: \"52f45a77-0faf-42ae-b052-ccb63b092440\") " pod="openstack/glance-db-create-27pv4" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.728776 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d1ed-account-create-update-64cxx" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.734792 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bclgq\" (UniqueName: \"kubernetes.io/projected/52f45a77-0faf-42ae-b052-ccb63b092440-kube-api-access-bclgq\") pod \"glance-db-create-27pv4\" (UID: \"52f45a77-0faf-42ae-b052-ccb63b092440\") " pod="openstack/glance-db-create-27pv4" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.785563 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ed22-account-create-update-c4lxx" event={"ID":"618420cf-adef-4233-b155-d213621354a0","Type":"ContainerStarted","Data":"06447f442fa1852213064c40ed55ae0f7ad4ed60dd89fe0aab732753b6816acf"} Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.803891 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-27pv4" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.818278 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5e17350-562f-4cd1-a861-0cea798c8528-operator-scripts\") pod \"glance-8170-account-create-update-8jz42\" (UID: \"d5e17350-562f-4cd1-a861-0cea798c8528\") " pod="openstack/glance-8170-account-create-update-8jz42" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.818333 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6rlx\" (UniqueName: \"kubernetes.io/projected/d5e17350-562f-4cd1-a861-0cea798c8528-kube-api-access-p6rlx\") pod \"glance-8170-account-create-update-8jz42\" (UID: \"d5e17350-562f-4cd1-a861-0cea798c8528\") " pod="openstack/glance-8170-account-create-update-8jz42" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.827268 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5e17350-562f-4cd1-a861-0cea798c8528-operator-scripts\") pod \"glance-8170-account-create-update-8jz42\" (UID: \"d5e17350-562f-4cd1-a861-0cea798c8528\") " pod="openstack/glance-8170-account-create-update-8jz42" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.828227 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-dxcwt"] Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.834584 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6rlx\" (UniqueName: \"kubernetes.io/projected/d5e17350-562f-4cd1-a861-0cea798c8528-kube-api-access-p6rlx\") pod \"glance-8170-account-create-update-8jz42\" (UID: \"d5e17350-562f-4cd1-a861-0cea798c8528\") " pod="openstack/glance-8170-account-create-update-8jz42" Dec 01 15:51:53 crc kubenswrapper[4739]: W1201 15:51:53.884507 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7ff093b_772b_4c09_ba41_6f5fd9324591.slice/crio-299fb3301d925efed8e5f43806eea385c914d6da71235322314eeaf031b13b25 WatchSource:0}: Error finding container 299fb3301d925efed8e5f43806eea385c914d6da71235322314eeaf031b13b25: Status 404 returned error can't find the container with id 299fb3301d925efed8e5f43806eea385c914d6da71235322314eeaf031b13b25 Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.911867 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.930035 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8170-account-create-update-8jz42" Dec 01 15:51:53 crc kubenswrapper[4739]: I1201 15:51:53.980918 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-skpk2"] Dec 01 15:51:53 crc kubenswrapper[4739]: W1201 15:51:53.985844 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9add3351_b3aa_41a1_8c08_5c10147e2520.slice/crio-2825191c676f3d5d8b6791a8edfdec0deca207c3c54c278167b3a93ffaf82e38 WatchSource:0}: Error finding container 2825191c676f3d5d8b6791a8edfdec0deca207c3c54c278167b3a93ffaf82e38: Status 404 returned error can't find the container with id 2825191c676f3d5d8b6791a8edfdec0deca207c3c54c278167b3a93ffaf82e38 Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.245624 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d1ed-account-create-update-64cxx"] Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.366135 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-27pv4"] Dec 01 15:51:54 crc kubenswrapper[4739]: W1201 15:51:54.424405 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52f45a77_0faf_42ae_b052_ccb63b092440.slice/crio-c02ca613a3f740d83017489098c81b606a928c83970f2624b3eaf8df67f5427e WatchSource:0}: Error finding container c02ca613a3f740d83017489098c81b606a928c83970f2624b3eaf8df67f5427e: Status 404 returned error can't find the container with id c02ca613a3f740d83017489098c81b606a928c83970f2624b3eaf8df67f5427e Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.473075 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-8170-account-create-update-8jz42"] Dec 01 15:51:54 crc kubenswrapper[4739]: W1201 15:51:54.532918 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5e17350_562f_4cd1_a861_0cea798c8528.slice/crio-6f1f6e6fa710db9d7d7130cd6199d94c98facc131aab9a9e9b12508a003b9a0f WatchSource:0}: Error finding container 6f1f6e6fa710db9d7d7130cd6199d94c98facc131aab9a9e9b12508a003b9a0f: Status 404 returned error can't find the container with id 6f1f6e6fa710db9d7d7130cd6199d94c98facc131aab9a9e9b12508a003b9a0f Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.797836 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8170-account-create-update-8jz42" event={"ID":"d5e17350-562f-4cd1-a861-0cea798c8528","Type":"ContainerStarted","Data":"1f66a2f72a31f5b7dc08bf035004c414c9c2d6d960dde14430a819e6c6ec1c38"} Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.798336 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8170-account-create-update-8jz42" event={"ID":"d5e17350-562f-4cd1-a861-0cea798c8528","Type":"ContainerStarted","Data":"6f1f6e6fa710db9d7d7130cd6199d94c98facc131aab9a9e9b12508a003b9a0f"} Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.799662 4739 generic.go:334] "Generic (PLEG): container finished" podID="9add3351-b3aa-41a1-8c08-5c10147e2520" containerID="d2f6a3cb0b76bdca0eab5cbaa0f63fe887ee181aa1f33abab0300196c6bc4c40" exitCode=0 Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.799849 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-skpk2" event={"ID":"9add3351-b3aa-41a1-8c08-5c10147e2520","Type":"ContainerDied","Data":"d2f6a3cb0b76bdca0eab5cbaa0f63fe887ee181aa1f33abab0300196c6bc4c40"} Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.799938 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-skpk2" event={"ID":"9add3351-b3aa-41a1-8c08-5c10147e2520","Type":"ContainerStarted","Data":"2825191c676f3d5d8b6791a8edfdec0deca207c3c54c278167b3a93ffaf82e38"} Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.803295 4739 generic.go:334] "Generic (PLEG): container finished" podID="f7ff093b-772b-4c09-ba41-6f5fd9324591" containerID="52a711a736dffeb17d2be406d7ef5ba5f2306a1e9a8afd8f83bc9db87ce8c990" exitCode=0 Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.803386 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-dxcwt" event={"ID":"f7ff093b-772b-4c09-ba41-6f5fd9324591","Type":"ContainerDied","Data":"52a711a736dffeb17d2be406d7ef5ba5f2306a1e9a8afd8f83bc9db87ce8c990"} Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.803522 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-dxcwt" event={"ID":"f7ff093b-772b-4c09-ba41-6f5fd9324591","Type":"ContainerStarted","Data":"299fb3301d925efed8e5f43806eea385c914d6da71235322314eeaf031b13b25"} Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.807449 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d1ed-account-create-update-64cxx" event={"ID":"dceccaf5-f087-477b-b9ac-4a0748136112","Type":"ContainerStarted","Data":"b18a60b2ae4f8812e5b38987deb886bf2bc8e2925c5afbc43170c982e5f56c09"} Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.807549 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d1ed-account-create-update-64cxx" event={"ID":"dceccaf5-f087-477b-b9ac-4a0748136112","Type":"ContainerStarted","Data":"2a6902cec74408c2fec28a13799bf463a3b83677b00b8fdc4acf1447ec132d9a"} Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.809069 4739 generic.go:334] "Generic (PLEG): container finished" podID="618420cf-adef-4233-b155-d213621354a0" containerID="d22e2824bf982944c902071f019940b381768974cc86e288e7fb4033793dc037" exitCode=0 Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.809285 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ed22-account-create-update-c4lxx" event={"ID":"618420cf-adef-4233-b155-d213621354a0","Type":"ContainerDied","Data":"d22e2824bf982944c902071f019940b381768974cc86e288e7fb4033793dc037"} Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.811458 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-27pv4" event={"ID":"52f45a77-0faf-42ae-b052-ccb63b092440","Type":"ContainerStarted","Data":"9eb8e3ce1a5302ca942a68401089eaede902fc7a45cee116ab78056a8aad1124"} Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.811546 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-27pv4" event={"ID":"52f45a77-0faf-42ae-b052-ccb63b092440","Type":"ContainerStarted","Data":"c02ca613a3f740d83017489098c81b606a928c83970f2624b3eaf8df67f5427e"} Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.821883 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-8170-account-create-update-8jz42" podStartSLOduration=1.8218607470000001 podStartE2EDuration="1.821860747s" podCreationTimestamp="2025-12-01 15:51:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:51:54.816948977 +0000 UTC m=+1016.642695111" watchObservedRunningTime="2025-12-01 15:51:54.821860747 +0000 UTC m=+1016.647606841" Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.848034 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-27pv4" podStartSLOduration=1.848003931 podStartE2EDuration="1.848003931s" podCreationTimestamp="2025-12-01 15:51:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:51:54.84637959 +0000 UTC m=+1016.672125714" watchObservedRunningTime="2025-12-01 15:51:54.848003931 +0000 UTC m=+1016.673750065" Dec 01 15:51:54 crc kubenswrapper[4739]: I1201 15:51:54.865025 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-d1ed-account-create-update-64cxx" podStartSLOduration=1.8649965530000001 podStartE2EDuration="1.864996553s" podCreationTimestamp="2025-12-01 15:51:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:51:54.861121723 +0000 UTC m=+1016.686867817" watchObservedRunningTime="2025-12-01 15:51:54.864996553 +0000 UTC m=+1016.690742687" Dec 01 15:51:55 crc kubenswrapper[4739]: I1201 15:51:55.825536 4739 generic.go:334] "Generic (PLEG): container finished" podID="52f45a77-0faf-42ae-b052-ccb63b092440" containerID="9eb8e3ce1a5302ca942a68401089eaede902fc7a45cee116ab78056a8aad1124" exitCode=0 Dec 01 15:51:55 crc kubenswrapper[4739]: I1201 15:51:55.826001 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-27pv4" event={"ID":"52f45a77-0faf-42ae-b052-ccb63b092440","Type":"ContainerDied","Data":"9eb8e3ce1a5302ca942a68401089eaede902fc7a45cee116ab78056a8aad1124"} Dec 01 15:51:55 crc kubenswrapper[4739]: I1201 15:51:55.828864 4739 generic.go:334] "Generic (PLEG): container finished" podID="d5e17350-562f-4cd1-a861-0cea798c8528" containerID="1f66a2f72a31f5b7dc08bf035004c414c9c2d6d960dde14430a819e6c6ec1c38" exitCode=0 Dec 01 15:51:55 crc kubenswrapper[4739]: I1201 15:51:55.829051 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8170-account-create-update-8jz42" event={"ID":"d5e17350-562f-4cd1-a861-0cea798c8528","Type":"ContainerDied","Data":"1f66a2f72a31f5b7dc08bf035004c414c9c2d6d960dde14430a819e6c6ec1c38"} Dec 01 15:51:55 crc kubenswrapper[4739]: I1201 15:51:55.833809 4739 generic.go:334] "Generic (PLEG): container finished" podID="dceccaf5-f087-477b-b9ac-4a0748136112" containerID="b18a60b2ae4f8812e5b38987deb886bf2bc8e2925c5afbc43170c982e5f56c09" exitCode=0 Dec 01 15:51:55 crc kubenswrapper[4739]: I1201 15:51:55.833904 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d1ed-account-create-update-64cxx" event={"ID":"dceccaf5-f087-477b-b9ac-4a0748136112","Type":"ContainerDied","Data":"b18a60b2ae4f8812e5b38987deb886bf2bc8e2925c5afbc43170c982e5f56c09"} Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.352741 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ed22-account-create-update-c4lxx" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.365905 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-skpk2" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.432066 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-dxcwt" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.529136 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9add3351-b3aa-41a1-8c08-5c10147e2520-operator-scripts\") pod \"9add3351-b3aa-41a1-8c08-5c10147e2520\" (UID: \"9add3351-b3aa-41a1-8c08-5c10147e2520\") " Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.529220 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7ff093b-772b-4c09-ba41-6f5fd9324591-operator-scripts\") pod \"f7ff093b-772b-4c09-ba41-6f5fd9324591\" (UID: \"f7ff093b-772b-4c09-ba41-6f5fd9324591\") " Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.529292 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcs6h\" (UniqueName: \"kubernetes.io/projected/618420cf-adef-4233-b155-d213621354a0-kube-api-access-kcs6h\") pod \"618420cf-adef-4233-b155-d213621354a0\" (UID: \"618420cf-adef-4233-b155-d213621354a0\") " Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.529453 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gptxv\" (UniqueName: \"kubernetes.io/projected/f7ff093b-772b-4c09-ba41-6f5fd9324591-kube-api-access-gptxv\") pod \"f7ff093b-772b-4c09-ba41-6f5fd9324591\" (UID: \"f7ff093b-772b-4c09-ba41-6f5fd9324591\") " Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.529553 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/618420cf-adef-4233-b155-d213621354a0-operator-scripts\") pod \"618420cf-adef-4233-b155-d213621354a0\" (UID: \"618420cf-adef-4233-b155-d213621354a0\") " Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.529643 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plvm2\" (UniqueName: \"kubernetes.io/projected/9add3351-b3aa-41a1-8c08-5c10147e2520-kube-api-access-plvm2\") pod \"9add3351-b3aa-41a1-8c08-5c10147e2520\" (UID: \"9add3351-b3aa-41a1-8c08-5c10147e2520\") " Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.530063 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9add3351-b3aa-41a1-8c08-5c10147e2520-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9add3351-b3aa-41a1-8c08-5c10147e2520" (UID: "9add3351-b3aa-41a1-8c08-5c10147e2520"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.530277 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9add3351-b3aa-41a1-8c08-5c10147e2520-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.530075 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7ff093b-772b-4c09-ba41-6f5fd9324591-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f7ff093b-772b-4c09-ba41-6f5fd9324591" (UID: "f7ff093b-772b-4c09-ba41-6f5fd9324591"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.530808 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/618420cf-adef-4233-b155-d213621354a0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "618420cf-adef-4233-b155-d213621354a0" (UID: "618420cf-adef-4233-b155-d213621354a0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.535263 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9add3351-b3aa-41a1-8c08-5c10147e2520-kube-api-access-plvm2" (OuterVolumeSpecName: "kube-api-access-plvm2") pod "9add3351-b3aa-41a1-8c08-5c10147e2520" (UID: "9add3351-b3aa-41a1-8c08-5c10147e2520"). InnerVolumeSpecName "kube-api-access-plvm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.537630 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7ff093b-772b-4c09-ba41-6f5fd9324591-kube-api-access-gptxv" (OuterVolumeSpecName: "kube-api-access-gptxv") pod "f7ff093b-772b-4c09-ba41-6f5fd9324591" (UID: "f7ff093b-772b-4c09-ba41-6f5fd9324591"). InnerVolumeSpecName "kube-api-access-gptxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.541877 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/618420cf-adef-4233-b155-d213621354a0-kube-api-access-kcs6h" (OuterVolumeSpecName: "kube-api-access-kcs6h") pod "618420cf-adef-4233-b155-d213621354a0" (UID: "618420cf-adef-4233-b155-d213621354a0"). InnerVolumeSpecName "kube-api-access-kcs6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.632620 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gptxv\" (UniqueName: \"kubernetes.io/projected/f7ff093b-772b-4c09-ba41-6f5fd9324591-kube-api-access-gptxv\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.632841 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/618420cf-adef-4233-b155-d213621354a0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.632858 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plvm2\" (UniqueName: \"kubernetes.io/projected/9add3351-b3aa-41a1-8c08-5c10147e2520-kube-api-access-plvm2\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.632870 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7ff093b-772b-4c09-ba41-6f5fd9324591-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.632883 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcs6h\" (UniqueName: \"kubernetes.io/projected/618420cf-adef-4233-b155-d213621354a0-kube-api-access-kcs6h\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.848555 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ed22-account-create-update-c4lxx" event={"ID":"618420cf-adef-4233-b155-d213621354a0","Type":"ContainerDied","Data":"06447f442fa1852213064c40ed55ae0f7ad4ed60dd89fe0aab732753b6816acf"} Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.848609 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06447f442fa1852213064c40ed55ae0f7ad4ed60dd89fe0aab732753b6816acf" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.848698 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ed22-account-create-update-c4lxx" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.851674 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-skpk2" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.851675 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-skpk2" event={"ID":"9add3351-b3aa-41a1-8c08-5c10147e2520","Type":"ContainerDied","Data":"2825191c676f3d5d8b6791a8edfdec0deca207c3c54c278167b3a93ffaf82e38"} Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.851849 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2825191c676f3d5d8b6791a8edfdec0deca207c3c54c278167b3a93ffaf82e38" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.853385 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-dxcwt" event={"ID":"f7ff093b-772b-4c09-ba41-6f5fd9324591","Type":"ContainerDied","Data":"299fb3301d925efed8e5f43806eea385c914d6da71235322314eeaf031b13b25"} Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.853470 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="299fb3301d925efed8e5f43806eea385c914d6da71235322314eeaf031b13b25" Dec 01 15:51:56 crc kubenswrapper[4739]: I1201 15:51:56.853526 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-dxcwt" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.196934 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8170-account-create-update-8jz42" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.274206 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d1ed-account-create-update-64cxx" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.340884 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-27pv4" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.348789 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6rlx\" (UniqueName: \"kubernetes.io/projected/d5e17350-562f-4cd1-a861-0cea798c8528-kube-api-access-p6rlx\") pod \"d5e17350-562f-4cd1-a861-0cea798c8528\" (UID: \"d5e17350-562f-4cd1-a861-0cea798c8528\") " Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.349056 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5e17350-562f-4cd1-a861-0cea798c8528-operator-scripts\") pod \"d5e17350-562f-4cd1-a861-0cea798c8528\" (UID: \"d5e17350-562f-4cd1-a861-0cea798c8528\") " Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.349901 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5e17350-562f-4cd1-a861-0cea798c8528-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d5e17350-562f-4cd1-a861-0cea798c8528" (UID: "d5e17350-562f-4cd1-a861-0cea798c8528"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.356290 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5e17350-562f-4cd1-a861-0cea798c8528-kube-api-access-p6rlx" (OuterVolumeSpecName: "kube-api-access-p6rlx") pod "d5e17350-562f-4cd1-a861-0cea798c8528" (UID: "d5e17350-562f-4cd1-a861-0cea798c8528"). InnerVolumeSpecName "kube-api-access-p6rlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.451443 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52f45a77-0faf-42ae-b052-ccb63b092440-operator-scripts\") pod \"52f45a77-0faf-42ae-b052-ccb63b092440\" (UID: \"52f45a77-0faf-42ae-b052-ccb63b092440\") " Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.451506 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxvw5\" (UniqueName: \"kubernetes.io/projected/dceccaf5-f087-477b-b9ac-4a0748136112-kube-api-access-qxvw5\") pod \"dceccaf5-f087-477b-b9ac-4a0748136112\" (UID: \"dceccaf5-f087-477b-b9ac-4a0748136112\") " Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.451587 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dceccaf5-f087-477b-b9ac-4a0748136112-operator-scripts\") pod \"dceccaf5-f087-477b-b9ac-4a0748136112\" (UID: \"dceccaf5-f087-477b-b9ac-4a0748136112\") " Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.451645 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bclgq\" (UniqueName: \"kubernetes.io/projected/52f45a77-0faf-42ae-b052-ccb63b092440-kube-api-access-bclgq\") pod \"52f45a77-0faf-42ae-b052-ccb63b092440\" (UID: \"52f45a77-0faf-42ae-b052-ccb63b092440\") " Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.452193 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6rlx\" (UniqueName: \"kubernetes.io/projected/d5e17350-562f-4cd1-a861-0cea798c8528-kube-api-access-p6rlx\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.452242 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5e17350-562f-4cd1-a861-0cea798c8528-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.452234 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52f45a77-0faf-42ae-b052-ccb63b092440-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "52f45a77-0faf-42ae-b052-ccb63b092440" (UID: "52f45a77-0faf-42ae-b052-ccb63b092440"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.452722 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dceccaf5-f087-477b-b9ac-4a0748136112-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dceccaf5-f087-477b-b9ac-4a0748136112" (UID: "dceccaf5-f087-477b-b9ac-4a0748136112"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.456482 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52f45a77-0faf-42ae-b052-ccb63b092440-kube-api-access-bclgq" (OuterVolumeSpecName: "kube-api-access-bclgq") pod "52f45a77-0faf-42ae-b052-ccb63b092440" (UID: "52f45a77-0faf-42ae-b052-ccb63b092440"). InnerVolumeSpecName "kube-api-access-bclgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.456585 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dceccaf5-f087-477b-b9ac-4a0748136112-kube-api-access-qxvw5" (OuterVolumeSpecName: "kube-api-access-qxvw5") pod "dceccaf5-f087-477b-b9ac-4a0748136112" (UID: "dceccaf5-f087-477b-b9ac-4a0748136112"). InnerVolumeSpecName "kube-api-access-qxvw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.554074 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dceccaf5-f087-477b-b9ac-4a0748136112-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.554110 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bclgq\" (UniqueName: \"kubernetes.io/projected/52f45a77-0faf-42ae-b052-ccb63b092440-kube-api-access-bclgq\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.554123 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52f45a77-0faf-42ae-b052-ccb63b092440-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.554133 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxvw5\" (UniqueName: \"kubernetes.io/projected/dceccaf5-f087-477b-b9ac-4a0748136112-kube-api-access-qxvw5\") on node \"crc\" DevicePath \"\"" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.866998 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8170-account-create-update-8jz42" event={"ID":"d5e17350-562f-4cd1-a861-0cea798c8528","Type":"ContainerDied","Data":"6f1f6e6fa710db9d7d7130cd6199d94c98facc131aab9a9e9b12508a003b9a0f"} Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.867044 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f1f6e6fa710db9d7d7130cd6199d94c98facc131aab9a9e9b12508a003b9a0f" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.867099 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8170-account-create-update-8jz42" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.869364 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d1ed-account-create-update-64cxx" event={"ID":"dceccaf5-f087-477b-b9ac-4a0748136112","Type":"ContainerDied","Data":"2a6902cec74408c2fec28a13799bf463a3b83677b00b8fdc4acf1447ec132d9a"} Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.869413 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a6902cec74408c2fec28a13799bf463a3b83677b00b8fdc4acf1447ec132d9a" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.869444 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d1ed-account-create-update-64cxx" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.872108 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-27pv4" event={"ID":"52f45a77-0faf-42ae-b052-ccb63b092440","Type":"ContainerDied","Data":"c02ca613a3f740d83017489098c81b606a928c83970f2624b3eaf8df67f5427e"} Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.872128 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c02ca613a3f740d83017489098c81b606a928c83970f2624b3eaf8df67f5427e" Dec 01 15:51:57 crc kubenswrapper[4739]: I1201 15:51:57.872166 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-27pv4" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.797915 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-8zdxf"] Dec 01 15:51:58 crc kubenswrapper[4739]: E1201 15:51:58.798976 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52f45a77-0faf-42ae-b052-ccb63b092440" containerName="mariadb-database-create" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.799021 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="52f45a77-0faf-42ae-b052-ccb63b092440" containerName="mariadb-database-create" Dec 01 15:51:58 crc kubenswrapper[4739]: E1201 15:51:58.799073 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5e17350-562f-4cd1-a861-0cea798c8528" containerName="mariadb-account-create-update" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.799091 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5e17350-562f-4cd1-a861-0cea798c8528" containerName="mariadb-account-create-update" Dec 01 15:51:58 crc kubenswrapper[4739]: E1201 15:51:58.799125 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ff093b-772b-4c09-ba41-6f5fd9324591" containerName="mariadb-database-create" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.799143 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ff093b-772b-4c09-ba41-6f5fd9324591" containerName="mariadb-database-create" Dec 01 15:51:58 crc kubenswrapper[4739]: E1201 15:51:58.799175 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dceccaf5-f087-477b-b9ac-4a0748136112" containerName="mariadb-account-create-update" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.799191 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="dceccaf5-f087-477b-b9ac-4a0748136112" containerName="mariadb-account-create-update" Dec 01 15:51:58 crc kubenswrapper[4739]: E1201 15:51:58.799241 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9add3351-b3aa-41a1-8c08-5c10147e2520" containerName="mariadb-database-create" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.799259 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="9add3351-b3aa-41a1-8c08-5c10147e2520" containerName="mariadb-database-create" Dec 01 15:51:58 crc kubenswrapper[4739]: E1201 15:51:58.799300 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="618420cf-adef-4233-b155-d213621354a0" containerName="mariadb-account-create-update" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.799317 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="618420cf-adef-4233-b155-d213621354a0" containerName="mariadb-account-create-update" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.799706 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="618420cf-adef-4233-b155-d213621354a0" containerName="mariadb-account-create-update" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.799732 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="dceccaf5-f087-477b-b9ac-4a0748136112" containerName="mariadb-account-create-update" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.799774 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ff093b-772b-4c09-ba41-6f5fd9324591" containerName="mariadb-database-create" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.799806 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="9add3351-b3aa-41a1-8c08-5c10147e2520" containerName="mariadb-database-create" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.799839 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="52f45a77-0faf-42ae-b052-ccb63b092440" containerName="mariadb-database-create" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.799870 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5e17350-562f-4cd1-a861-0cea798c8528" containerName="mariadb-account-create-update" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.800980 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8zdxf" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.804519 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-src6f" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.804518 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.812924 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8zdxf"] Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.974513 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-db-sync-config-data\") pod \"glance-db-sync-8zdxf\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " pod="openstack/glance-db-sync-8zdxf" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.974608 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvrpj\" (UniqueName: \"kubernetes.io/projected/bc6eb674-5455-4596-8a2a-5b562dd87ef3-kube-api-access-mvrpj\") pod \"glance-db-sync-8zdxf\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " pod="openstack/glance-db-sync-8zdxf" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.974698 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-config-data\") pod \"glance-db-sync-8zdxf\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " pod="openstack/glance-db-sync-8zdxf" Dec 01 15:51:58 crc kubenswrapper[4739]: I1201 15:51:58.974900 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-combined-ca-bundle\") pod \"glance-db-sync-8zdxf\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " pod="openstack/glance-db-sync-8zdxf" Dec 01 15:51:59 crc kubenswrapper[4739]: I1201 15:51:59.077487 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvrpj\" (UniqueName: \"kubernetes.io/projected/bc6eb674-5455-4596-8a2a-5b562dd87ef3-kube-api-access-mvrpj\") pod \"glance-db-sync-8zdxf\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " pod="openstack/glance-db-sync-8zdxf" Dec 01 15:51:59 crc kubenswrapper[4739]: I1201 15:51:59.077764 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-config-data\") pod \"glance-db-sync-8zdxf\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " pod="openstack/glance-db-sync-8zdxf" Dec 01 15:51:59 crc kubenswrapper[4739]: I1201 15:51:59.077900 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-combined-ca-bundle\") pod \"glance-db-sync-8zdxf\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " pod="openstack/glance-db-sync-8zdxf" Dec 01 15:51:59 crc kubenswrapper[4739]: I1201 15:51:59.077989 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-db-sync-config-data\") pod \"glance-db-sync-8zdxf\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " pod="openstack/glance-db-sync-8zdxf" Dec 01 15:51:59 crc kubenswrapper[4739]: I1201 15:51:59.085376 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-combined-ca-bundle\") pod \"glance-db-sync-8zdxf\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " pod="openstack/glance-db-sync-8zdxf" Dec 01 15:51:59 crc kubenswrapper[4739]: I1201 15:51:59.085679 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-config-data\") pod \"glance-db-sync-8zdxf\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " pod="openstack/glance-db-sync-8zdxf" Dec 01 15:51:59 crc kubenswrapper[4739]: I1201 15:51:59.088014 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-db-sync-config-data\") pod \"glance-db-sync-8zdxf\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " pod="openstack/glance-db-sync-8zdxf" Dec 01 15:51:59 crc kubenswrapper[4739]: I1201 15:51:59.109349 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvrpj\" (UniqueName: \"kubernetes.io/projected/bc6eb674-5455-4596-8a2a-5b562dd87ef3-kube-api-access-mvrpj\") pod \"glance-db-sync-8zdxf\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " pod="openstack/glance-db-sync-8zdxf" Dec 01 15:51:59 crc kubenswrapper[4739]: I1201 15:51:59.135985 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8zdxf" Dec 01 15:51:59 crc kubenswrapper[4739]: I1201 15:51:59.712919 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8zdxf"] Dec 01 15:51:59 crc kubenswrapper[4739]: I1201 15:51:59.896384 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8zdxf" event={"ID":"bc6eb674-5455-4596-8a2a-5b562dd87ef3","Type":"ContainerStarted","Data":"455e69cd70d28ac423e72b8419cc7a39047a82246e36a4c740b63d32b164cf25"} Dec 01 15:52:00 crc kubenswrapper[4739]: I1201 15:52:00.449924 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 01 15:52:13 crc kubenswrapper[4739]: I1201 15:52:13.028442 4739 generic.go:334] "Generic (PLEG): container finished" podID="a2606100-66af-4acd-98a5-2edcfb281544" containerID="184f65dcda1a366c8645ddf60dcc32681e541fb0943f2bcdab6db7840d07555c" exitCode=0 Dec 01 15:52:13 crc kubenswrapper[4739]: I1201 15:52:13.028610 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a2606100-66af-4acd-98a5-2edcfb281544","Type":"ContainerDied","Data":"184f65dcda1a366c8645ddf60dcc32681e541fb0943f2bcdab6db7840d07555c"} Dec 01 15:52:14 crc kubenswrapper[4739]: I1201 15:52:14.037892 4739 generic.go:334] "Generic (PLEG): container finished" podID="0f128863-acb7-4080-829a-01bd78766374" containerID="17bbc7387e706b89896707fd45f1fbe1e75abe86f809cbdf894a0b9ec75c4520" exitCode=0 Dec 01 15:52:14 crc kubenswrapper[4739]: I1201 15:52:14.038013 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f128863-acb7-4080-829a-01bd78766374","Type":"ContainerDied","Data":"17bbc7387e706b89896707fd45f1fbe1e75abe86f809cbdf894a0b9ec75c4520"} Dec 01 15:52:14 crc kubenswrapper[4739]: I1201 15:52:14.042483 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a2606100-66af-4acd-98a5-2edcfb281544","Type":"ContainerStarted","Data":"44d5f06f08742ef999ca64710212d1783cbaa21433efd216736620fff2acdb03"} Dec 01 15:52:14 crc kubenswrapper[4739]: I1201 15:52:14.042726 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:52:14 crc kubenswrapper[4739]: I1201 15:52:14.045349 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8zdxf" event={"ID":"bc6eb674-5455-4596-8a2a-5b562dd87ef3","Type":"ContainerStarted","Data":"d1b6b216abfc000d4755183b3e4f3b04155939bb36df24644d4bf8bdf1abfaa6"} Dec 01 15:52:14 crc kubenswrapper[4739]: I1201 15:52:14.110583 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-8zdxf" podStartSLOduration=3.093357076 podStartE2EDuration="16.110563672s" podCreationTimestamp="2025-12-01 15:51:58 +0000 UTC" firstStartedPulling="2025-12-01 15:51:59.71531958 +0000 UTC m=+1021.541065704" lastFinishedPulling="2025-12-01 15:52:12.732526206 +0000 UTC m=+1034.558272300" observedRunningTime="2025-12-01 15:52:14.10268351 +0000 UTC m=+1035.928429624" watchObservedRunningTime="2025-12-01 15:52:14.110563672 +0000 UTC m=+1035.936309766" Dec 01 15:52:14 crc kubenswrapper[4739]: I1201 15:52:14.135528 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=54.232587084 podStartE2EDuration="1m6.135505058s" podCreationTimestamp="2025-12-01 15:51:08 +0000 UTC" firstStartedPulling="2025-12-01 15:51:24.836823907 +0000 UTC m=+986.662570001" lastFinishedPulling="2025-12-01 15:51:36.739741871 +0000 UTC m=+998.565487975" observedRunningTime="2025-12-01 15:52:14.131143744 +0000 UTC m=+1035.956889848" watchObservedRunningTime="2025-12-01 15:52:14.135505058 +0000 UTC m=+1035.961251162" Dec 01 15:52:14 crc kubenswrapper[4739]: I1201 15:52:14.680814 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4ltrp" podUID="10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15" containerName="ovn-controller" probeResult="failure" output=< Dec 01 15:52:14 crc kubenswrapper[4739]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 01 15:52:14 crc kubenswrapper[4739]: > Dec 01 15:52:14 crc kubenswrapper[4739]: I1201 15:52:14.693044 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:52:14 crc kubenswrapper[4739]: I1201 15:52:14.713685 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-pr8vl" Dec 01 15:52:14 crc kubenswrapper[4739]: I1201 15:52:14.923778 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4ltrp-config-d5ndt"] Dec 01 15:52:14 crc kubenswrapper[4739]: I1201 15:52:14.924706 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:14 crc kubenswrapper[4739]: I1201 15:52:14.927286 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 01 15:52:14 crc kubenswrapper[4739]: I1201 15:52:14.951923 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ltrp-config-d5ndt"] Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.029749 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvp6h\" (UniqueName: \"kubernetes.io/projected/cd6c8a32-e45e-4476-b046-5f36471a1cdc-kube-api-access-tvp6h\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.030465 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd6c8a32-e45e-4476-b046-5f36471a1cdc-scripts\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.030572 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-run\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.030701 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-log-ovn\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.030824 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cd6c8a32-e45e-4476-b046-5f36471a1cdc-additional-scripts\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.030940 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-run-ovn\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.058579 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f128863-acb7-4080-829a-01bd78766374","Type":"ContainerStarted","Data":"40c19226330b8e279fe364bce8b752d5af236a3696e9b11962cef197f1425812"} Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.061498 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.084466 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=55.25228211 podStartE2EDuration="1m7.08444024s" podCreationTimestamp="2025-12-01 15:51:08 +0000 UTC" firstStartedPulling="2025-12-01 15:51:24.206375238 +0000 UTC m=+986.032121332" lastFinishedPulling="2025-12-01 15:51:36.038533358 +0000 UTC m=+997.864279462" observedRunningTime="2025-12-01 15:52:15.080193 +0000 UTC m=+1036.905939134" watchObservedRunningTime="2025-12-01 15:52:15.08444024 +0000 UTC m=+1036.910186374" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.133063 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd6c8a32-e45e-4476-b046-5f36471a1cdc-scripts\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.133482 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-run\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.133618 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-log-ovn\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.133830 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cd6c8a32-e45e-4476-b046-5f36471a1cdc-additional-scripts\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.133914 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-log-ovn\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.134095 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-run-ovn\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.134211 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-run-ovn\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.134453 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvp6h\" (UniqueName: \"kubernetes.io/projected/cd6c8a32-e45e-4476-b046-5f36471a1cdc-kube-api-access-tvp6h\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.134683 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cd6c8a32-e45e-4476-b046-5f36471a1cdc-additional-scripts\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.134751 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-run\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.135738 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd6c8a32-e45e-4476-b046-5f36471a1cdc-scripts\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.163387 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvp6h\" (UniqueName: \"kubernetes.io/projected/cd6c8a32-e45e-4476-b046-5f36471a1cdc-kube-api-access-tvp6h\") pod \"ovn-controller-4ltrp-config-d5ndt\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.250879 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:15 crc kubenswrapper[4739]: I1201 15:52:15.795790 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ltrp-config-d5ndt"] Dec 01 15:52:15 crc kubenswrapper[4739]: W1201 15:52:15.801647 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd6c8a32_e45e_4476_b046_5f36471a1cdc.slice/crio-ae10b8b9cbd15de837730ab2d96032ecba79fde0fe0b6180441a1affa5d2612c WatchSource:0}: Error finding container ae10b8b9cbd15de837730ab2d96032ecba79fde0fe0b6180441a1affa5d2612c: Status 404 returned error can't find the container with id ae10b8b9cbd15de837730ab2d96032ecba79fde0fe0b6180441a1affa5d2612c Dec 01 15:52:16 crc kubenswrapper[4739]: I1201 15:52:16.067145 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ltrp-config-d5ndt" event={"ID":"cd6c8a32-e45e-4476-b046-5f36471a1cdc","Type":"ContainerStarted","Data":"ae10b8b9cbd15de837730ab2d96032ecba79fde0fe0b6180441a1affa5d2612c"} Dec 01 15:52:17 crc kubenswrapper[4739]: I1201 15:52:17.080957 4739 generic.go:334] "Generic (PLEG): container finished" podID="cd6c8a32-e45e-4476-b046-5f36471a1cdc" containerID="7c3c3f663d2c893ed61b1ce9e3eb2b38720420458e5eb4a3a399b87cb2d2c964" exitCode=0 Dec 01 15:52:17 crc kubenswrapper[4739]: I1201 15:52:17.081046 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ltrp-config-d5ndt" event={"ID":"cd6c8a32-e45e-4476-b046-5f36471a1cdc","Type":"ContainerDied","Data":"7c3c3f663d2c893ed61b1ce9e3eb2b38720420458e5eb4a3a399b87cb2d2c964"} Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.378196 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.488854 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cd6c8a32-e45e-4476-b046-5f36471a1cdc-additional-scripts\") pod \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.488947 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-log-ovn\") pod \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.489012 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-run-ovn\") pod \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.489059 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "cd6c8a32-e45e-4476-b046-5f36471a1cdc" (UID: "cd6c8a32-e45e-4476-b046-5f36471a1cdc"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.489103 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvp6h\" (UniqueName: \"kubernetes.io/projected/cd6c8a32-e45e-4476-b046-5f36471a1cdc-kube-api-access-tvp6h\") pod \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.489123 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "cd6c8a32-e45e-4476-b046-5f36471a1cdc" (UID: "cd6c8a32-e45e-4476-b046-5f36471a1cdc"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.489129 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd6c8a32-e45e-4476-b046-5f36471a1cdc-scripts\") pod \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.489223 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-run\") pod \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\" (UID: \"cd6c8a32-e45e-4476-b046-5f36471a1cdc\") " Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.489321 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-run" (OuterVolumeSpecName: "var-run") pod "cd6c8a32-e45e-4476-b046-5f36471a1cdc" (UID: "cd6c8a32-e45e-4476-b046-5f36471a1cdc"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.489738 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd6c8a32-e45e-4476-b046-5f36471a1cdc-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "cd6c8a32-e45e-4476-b046-5f36471a1cdc" (UID: "cd6c8a32-e45e-4476-b046-5f36471a1cdc"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.489949 4739 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-run\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.489966 4739 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cd6c8a32-e45e-4476-b046-5f36471a1cdc-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.489981 4739 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.489993 4739 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd6c8a32-e45e-4476-b046-5f36471a1cdc-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.490104 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd6c8a32-e45e-4476-b046-5f36471a1cdc-scripts" (OuterVolumeSpecName: "scripts") pod "cd6c8a32-e45e-4476-b046-5f36471a1cdc" (UID: "cd6c8a32-e45e-4476-b046-5f36471a1cdc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.497533 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd6c8a32-e45e-4476-b046-5f36471a1cdc-kube-api-access-tvp6h" (OuterVolumeSpecName: "kube-api-access-tvp6h") pod "cd6c8a32-e45e-4476-b046-5f36471a1cdc" (UID: "cd6c8a32-e45e-4476-b046-5f36471a1cdc"). InnerVolumeSpecName "kube-api-access-tvp6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.591731 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvp6h\" (UniqueName: \"kubernetes.io/projected/cd6c8a32-e45e-4476-b046-5f36471a1cdc-kube-api-access-tvp6h\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:18 crc kubenswrapper[4739]: I1201 15:52:18.591920 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd6c8a32-e45e-4476-b046-5f36471a1cdc-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.096035 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ltrp-config-d5ndt" event={"ID":"cd6c8a32-e45e-4476-b046-5f36471a1cdc","Type":"ContainerDied","Data":"ae10b8b9cbd15de837730ab2d96032ecba79fde0fe0b6180441a1affa5d2612c"} Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.096069 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae10b8b9cbd15de837730ab2d96032ecba79fde0fe0b6180441a1affa5d2612c" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.096147 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ltrp-config-d5ndt" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.526667 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4ltrp-config-d5ndt"] Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.534390 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4ltrp-config-d5ndt"] Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.590448 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4ltrp-config-89kx4"] Dec 01 15:52:19 crc kubenswrapper[4739]: E1201 15:52:19.590944 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd6c8a32-e45e-4476-b046-5f36471a1cdc" containerName="ovn-config" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.590972 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd6c8a32-e45e-4476-b046-5f36471a1cdc" containerName="ovn-config" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.591225 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd6c8a32-e45e-4476-b046-5f36471a1cdc" containerName="ovn-config" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.592131 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.594272 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.608107 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ltrp-config-89kx4"] Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.689656 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-4ltrp" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.717578 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq4bc\" (UniqueName: \"kubernetes.io/projected/38af4519-4415-4f20-b447-4177056cd960-kube-api-access-vq4bc\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.717690 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-log-ovn\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.717863 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-run-ovn\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.717911 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-run\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.717961 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38af4519-4415-4f20-b447-4177056cd960-scripts\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.717991 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/38af4519-4415-4f20-b447-4177056cd960-additional-scripts\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.819345 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq4bc\" (UniqueName: \"kubernetes.io/projected/38af4519-4415-4f20-b447-4177056cd960-kube-api-access-vq4bc\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.819398 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-log-ovn\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.819536 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-run-ovn\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.819608 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-run\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.819628 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38af4519-4415-4f20-b447-4177056cd960-scripts\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.819653 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/38af4519-4415-4f20-b447-4177056cd960-additional-scripts\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.819847 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-log-ovn\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.820112 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-run-ovn\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.820538 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-run\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.820694 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/38af4519-4415-4f20-b447-4177056cd960-additional-scripts\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.821826 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38af4519-4415-4f20-b447-4177056cd960-scripts\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.848261 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq4bc\" (UniqueName: \"kubernetes.io/projected/38af4519-4415-4f20-b447-4177056cd960-kube-api-access-vq4bc\") pod \"ovn-controller-4ltrp-config-89kx4\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:19 crc kubenswrapper[4739]: I1201 15:52:19.911001 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:20 crc kubenswrapper[4739]: I1201 15:52:20.112973 4739 generic.go:334] "Generic (PLEG): container finished" podID="bc6eb674-5455-4596-8a2a-5b562dd87ef3" containerID="d1b6b216abfc000d4755183b3e4f3b04155939bb36df24644d4bf8bdf1abfaa6" exitCode=0 Dec 01 15:52:20 crc kubenswrapper[4739]: I1201 15:52:20.113309 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8zdxf" event={"ID":"bc6eb674-5455-4596-8a2a-5b562dd87ef3","Type":"ContainerDied","Data":"d1b6b216abfc000d4755183b3e4f3b04155939bb36df24644d4bf8bdf1abfaa6"} Dec 01 15:52:20 crc kubenswrapper[4739]: I1201 15:52:20.177271 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ltrp-config-89kx4"] Dec 01 15:52:20 crc kubenswrapper[4739]: W1201 15:52:20.181740 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38af4519_4415_4f20_b447_4177056cd960.slice/crio-95f33db4c418cc8ef7cac507ebf7037ead0cee1873a6ebab8df83974b52590ea WatchSource:0}: Error finding container 95f33db4c418cc8ef7cac507ebf7037ead0cee1873a6ebab8df83974b52590ea: Status 404 returned error can't find the container with id 95f33db4c418cc8ef7cac507ebf7037ead0cee1873a6ebab8df83974b52590ea Dec 01 15:52:20 crc kubenswrapper[4739]: I1201 15:52:20.485046 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd6c8a32-e45e-4476-b046-5f36471a1cdc" path="/var/lib/kubelet/pods/cd6c8a32-e45e-4476-b046-5f36471a1cdc/volumes" Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.120976 4739 generic.go:334] "Generic (PLEG): container finished" podID="38af4519-4415-4f20-b447-4177056cd960" containerID="d0f94041a898118ba56ed8b833fda3241a2179e10a73ae2ca4a679e00f5d0128" exitCode=0 Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.121023 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ltrp-config-89kx4" event={"ID":"38af4519-4415-4f20-b447-4177056cd960","Type":"ContainerDied","Data":"d0f94041a898118ba56ed8b833fda3241a2179e10a73ae2ca4a679e00f5d0128"} Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.121290 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ltrp-config-89kx4" event={"ID":"38af4519-4415-4f20-b447-4177056cd960","Type":"ContainerStarted","Data":"95f33db4c418cc8ef7cac507ebf7037ead0cee1873a6ebab8df83974b52590ea"} Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.504929 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8zdxf" Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.576608 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvrpj\" (UniqueName: \"kubernetes.io/projected/bc6eb674-5455-4596-8a2a-5b562dd87ef3-kube-api-access-mvrpj\") pod \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.576662 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-db-sync-config-data\") pod \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.576717 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-config-data\") pod \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.576771 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-combined-ca-bundle\") pod \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\" (UID: \"bc6eb674-5455-4596-8a2a-5b562dd87ef3\") " Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.595047 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc6eb674-5455-4596-8a2a-5b562dd87ef3-kube-api-access-mvrpj" (OuterVolumeSpecName: "kube-api-access-mvrpj") pod "bc6eb674-5455-4596-8a2a-5b562dd87ef3" (UID: "bc6eb674-5455-4596-8a2a-5b562dd87ef3"). InnerVolumeSpecName "kube-api-access-mvrpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.595209 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bc6eb674-5455-4596-8a2a-5b562dd87ef3" (UID: "bc6eb674-5455-4596-8a2a-5b562dd87ef3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.615690 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc6eb674-5455-4596-8a2a-5b562dd87ef3" (UID: "bc6eb674-5455-4596-8a2a-5b562dd87ef3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.638744 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-config-data" (OuterVolumeSpecName: "config-data") pod "bc6eb674-5455-4596-8a2a-5b562dd87ef3" (UID: "bc6eb674-5455-4596-8a2a-5b562dd87ef3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.678751 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.678779 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvrpj\" (UniqueName: \"kubernetes.io/projected/bc6eb674-5455-4596-8a2a-5b562dd87ef3-kube-api-access-mvrpj\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.678789 4739 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:21 crc kubenswrapper[4739]: I1201 15:52:21.678799 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc6eb674-5455-4596-8a2a-5b562dd87ef3-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.133498 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8zdxf" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.133506 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8zdxf" event={"ID":"bc6eb674-5455-4596-8a2a-5b562dd87ef3","Type":"ContainerDied","Data":"455e69cd70d28ac423e72b8419cc7a39047a82246e36a4c740b63d32b164cf25"} Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.133570 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="455e69cd70d28ac423e72b8419cc7a39047a82246e36a4c740b63d32b164cf25" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.553318 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.583754 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-nqkgc"] Dec 01 15:52:22 crc kubenswrapper[4739]: E1201 15:52:22.584176 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc6eb674-5455-4596-8a2a-5b562dd87ef3" containerName="glance-db-sync" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.584194 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc6eb674-5455-4596-8a2a-5b562dd87ef3" containerName="glance-db-sync" Dec 01 15:52:22 crc kubenswrapper[4739]: E1201 15:52:22.584209 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38af4519-4415-4f20-b447-4177056cd960" containerName="ovn-config" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.584216 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="38af4519-4415-4f20-b447-4177056cd960" containerName="ovn-config" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.584352 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="38af4519-4415-4f20-b447-4177056cd960" containerName="ovn-config" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.584371 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc6eb674-5455-4596-8a2a-5b562dd87ef3" containerName="glance-db-sync" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.585219 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.599953 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38af4519-4415-4f20-b447-4177056cd960-scripts\") pod \"38af4519-4415-4f20-b447-4177056cd960\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.600008 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-run-ovn\") pod \"38af4519-4415-4f20-b447-4177056cd960\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.600070 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-log-ovn\") pod \"38af4519-4415-4f20-b447-4177056cd960\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.600094 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/38af4519-4415-4f20-b447-4177056cd960-additional-scripts\") pod \"38af4519-4415-4f20-b447-4177056cd960\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.600127 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vq4bc\" (UniqueName: \"kubernetes.io/projected/38af4519-4415-4f20-b447-4177056cd960-kube-api-access-vq4bc\") pod \"38af4519-4415-4f20-b447-4177056cd960\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.600147 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-run\") pod \"38af4519-4415-4f20-b447-4177056cd960\" (UID: \"38af4519-4415-4f20-b447-4177056cd960\") " Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.600830 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-ovsdbserver-nb\") pod \"dnsmasq-dns-54f9b7b8d9-nqkgc\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.600868 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-ovsdbserver-sb\") pod \"dnsmasq-dns-54f9b7b8d9-nqkgc\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.600886 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-dns-svc\") pod \"dnsmasq-dns-54f9b7b8d9-nqkgc\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.600904 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-config\") pod \"dnsmasq-dns-54f9b7b8d9-nqkgc\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.600933 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxr7t\" (UniqueName: \"kubernetes.io/projected/9ffc2832-c824-470f-8cfa-9ad4ea814125-kube-api-access-rxr7t\") pod \"dnsmasq-dns-54f9b7b8d9-nqkgc\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.600967 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "38af4519-4415-4f20-b447-4177056cd960" (UID: "38af4519-4415-4f20-b447-4177056cd960"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.601287 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38af4519-4415-4f20-b447-4177056cd960-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "38af4519-4415-4f20-b447-4177056cd960" (UID: "38af4519-4415-4f20-b447-4177056cd960"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.601744 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38af4519-4415-4f20-b447-4177056cd960-scripts" (OuterVolumeSpecName: "scripts") pod "38af4519-4415-4f20-b447-4177056cd960" (UID: "38af4519-4415-4f20-b447-4177056cd960"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.601781 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "38af4519-4415-4f20-b447-4177056cd960" (UID: "38af4519-4415-4f20-b447-4177056cd960"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.601810 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-run" (OuterVolumeSpecName: "var-run") pod "38af4519-4415-4f20-b447-4177056cd960" (UID: "38af4519-4415-4f20-b447-4177056cd960"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.605778 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38af4519-4415-4f20-b447-4177056cd960-kube-api-access-vq4bc" (OuterVolumeSpecName: "kube-api-access-vq4bc") pod "38af4519-4415-4f20-b447-4177056cd960" (UID: "38af4519-4415-4f20-b447-4177056cd960"). InnerVolumeSpecName "kube-api-access-vq4bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.620222 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-nqkgc"] Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.702468 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-ovsdbserver-nb\") pod \"dnsmasq-dns-54f9b7b8d9-nqkgc\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.702765 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-ovsdbserver-sb\") pod \"dnsmasq-dns-54f9b7b8d9-nqkgc\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.702786 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-dns-svc\") pod \"dnsmasq-dns-54f9b7b8d9-nqkgc\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.702808 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-config\") pod \"dnsmasq-dns-54f9b7b8d9-nqkgc\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.702837 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxr7t\" (UniqueName: \"kubernetes.io/projected/9ffc2832-c824-470f-8cfa-9ad4ea814125-kube-api-access-rxr7t\") pod \"dnsmasq-dns-54f9b7b8d9-nqkgc\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.702878 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38af4519-4415-4f20-b447-4177056cd960-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.702888 4739 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.702896 4739 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.702906 4739 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/38af4519-4415-4f20-b447-4177056cd960-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.702915 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vq4bc\" (UniqueName: \"kubernetes.io/projected/38af4519-4415-4f20-b447-4177056cd960-kube-api-access-vq4bc\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.702923 4739 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/38af4519-4415-4f20-b447-4177056cd960-var-run\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.703916 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-ovsdbserver-nb\") pod \"dnsmasq-dns-54f9b7b8d9-nqkgc\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.704093 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-dns-svc\") pod \"dnsmasq-dns-54f9b7b8d9-nqkgc\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.704310 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-ovsdbserver-sb\") pod \"dnsmasq-dns-54f9b7b8d9-nqkgc\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.704857 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-config\") pod \"dnsmasq-dns-54f9b7b8d9-nqkgc\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.717328 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxr7t\" (UniqueName: \"kubernetes.io/projected/9ffc2832-c824-470f-8cfa-9ad4ea814125-kube-api-access-rxr7t\") pod \"dnsmasq-dns-54f9b7b8d9-nqkgc\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:22 crc kubenswrapper[4739]: I1201 15:52:22.902370 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:23 crc kubenswrapper[4739]: I1201 15:52:23.144396 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ltrp-config-89kx4" event={"ID":"38af4519-4415-4f20-b447-4177056cd960","Type":"ContainerDied","Data":"95f33db4c418cc8ef7cac507ebf7037ead0cee1873a6ebab8df83974b52590ea"} Dec 01 15:52:23 crc kubenswrapper[4739]: I1201 15:52:23.144868 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95f33db4c418cc8ef7cac507ebf7037ead0cee1873a6ebab8df83974b52590ea" Dec 01 15:52:23 crc kubenswrapper[4739]: I1201 15:52:23.144699 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ltrp-config-89kx4" Dec 01 15:52:23 crc kubenswrapper[4739]: I1201 15:52:23.334677 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-nqkgc"] Dec 01 15:52:23 crc kubenswrapper[4739]: I1201 15:52:23.655743 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4ltrp-config-89kx4"] Dec 01 15:52:23 crc kubenswrapper[4739]: I1201 15:52:23.661286 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4ltrp-config-89kx4"] Dec 01 15:52:24 crc kubenswrapper[4739]: I1201 15:52:24.155595 4739 generic.go:334] "Generic (PLEG): container finished" podID="9ffc2832-c824-470f-8cfa-9ad4ea814125" containerID="974146f86ecb73d73747c3e21331e32fe0162b0168be30b462ed1f8dcc308199" exitCode=0 Dec 01 15:52:24 crc kubenswrapper[4739]: I1201 15:52:24.155838 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" event={"ID":"9ffc2832-c824-470f-8cfa-9ad4ea814125","Type":"ContainerDied","Data":"974146f86ecb73d73747c3e21331e32fe0162b0168be30b462ed1f8dcc308199"} Dec 01 15:52:24 crc kubenswrapper[4739]: I1201 15:52:24.155861 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" event={"ID":"9ffc2832-c824-470f-8cfa-9ad4ea814125","Type":"ContainerStarted","Data":"a1db62b465838bf05e97955b72b236b49b13c059eeac9c511e2db4afd0f8c49e"} Dec 01 15:52:24 crc kubenswrapper[4739]: I1201 15:52:24.491280 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38af4519-4415-4f20-b447-4177056cd960" path="/var/lib/kubelet/pods/38af4519-4415-4f20-b447-4177056cd960/volumes" Dec 01 15:52:25 crc kubenswrapper[4739]: I1201 15:52:25.168963 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" event={"ID":"9ffc2832-c824-470f-8cfa-9ad4ea814125","Type":"ContainerStarted","Data":"9386ffe9422d13a240a8f5e843765882e844a274968e8d74e834de01f6d2566c"} Dec 01 15:52:25 crc kubenswrapper[4739]: I1201 15:52:25.169447 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:25 crc kubenswrapper[4739]: I1201 15:52:25.196796 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" podStartSLOduration=3.196767979 podStartE2EDuration="3.196767979s" podCreationTimestamp="2025-12-01 15:52:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:52:25.196642765 +0000 UTC m=+1047.022388869" watchObservedRunningTime="2025-12-01 15:52:25.196767979 +0000 UTC m=+1047.022514113" Dec 01 15:52:29 crc kubenswrapper[4739]: I1201 15:52:29.896673 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.195841 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.433630 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-6sn86"] Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.438652 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6sn86" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.444114 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6sn86"] Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.464407 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbzq5\" (UniqueName: \"kubernetes.io/projected/56bbe10e-0b3e-48c5-a806-cf5f1dec0935-kube-api-access-lbzq5\") pod \"cinder-db-create-6sn86\" (UID: \"56bbe10e-0b3e-48c5-a806-cf5f1dec0935\") " pod="openstack/cinder-db-create-6sn86" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.464539 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56bbe10e-0b3e-48c5-a806-cf5f1dec0935-operator-scripts\") pod \"cinder-db-create-6sn86\" (UID: \"56bbe10e-0b3e-48c5-a806-cf5f1dec0935\") " pod="openstack/cinder-db-create-6sn86" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.565492 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56bbe10e-0b3e-48c5-a806-cf5f1dec0935-operator-scripts\") pod \"cinder-db-create-6sn86\" (UID: \"56bbe10e-0b3e-48c5-a806-cf5f1dec0935\") " pod="openstack/cinder-db-create-6sn86" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.565567 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbzq5\" (UniqueName: \"kubernetes.io/projected/56bbe10e-0b3e-48c5-a806-cf5f1dec0935-kube-api-access-lbzq5\") pod \"cinder-db-create-6sn86\" (UID: \"56bbe10e-0b3e-48c5-a806-cf5f1dec0935\") " pod="openstack/cinder-db-create-6sn86" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.566863 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56bbe10e-0b3e-48c5-a806-cf5f1dec0935-operator-scripts\") pod \"cinder-db-create-6sn86\" (UID: \"56bbe10e-0b3e-48c5-a806-cf5f1dec0935\") " pod="openstack/cinder-db-create-6sn86" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.628842 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbzq5\" (UniqueName: \"kubernetes.io/projected/56bbe10e-0b3e-48c5-a806-cf5f1dec0935-kube-api-access-lbzq5\") pod \"cinder-db-create-6sn86\" (UID: \"56bbe10e-0b3e-48c5-a806-cf5f1dec0935\") " pod="openstack/cinder-db-create-6sn86" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.640496 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-4jbjk"] Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.648671 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4jbjk" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.670895 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b6d7\" (UniqueName: \"kubernetes.io/projected/58e52c5d-7e41-4145-abad-945e9cb9d01b-kube-api-access-6b6d7\") pod \"barbican-db-create-4jbjk\" (UID: \"58e52c5d-7e41-4145-abad-945e9cb9d01b\") " pod="openstack/barbican-db-create-4jbjk" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.671325 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58e52c5d-7e41-4145-abad-945e9cb9d01b-operator-scripts\") pod \"barbican-db-create-4jbjk\" (UID: \"58e52c5d-7e41-4145-abad-945e9cb9d01b\") " pod="openstack/barbican-db-create-4jbjk" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.673411 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4jbjk"] Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.682812 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-1814-account-create-update-7l5hz"] Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.683956 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1814-account-create-update-7l5hz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.689087 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.706817 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1814-account-create-update-7l5hz"] Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.714521 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-fm2qz"] Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.715501 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-fm2qz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.723130 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.723240 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zcbnm" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.723336 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.723533 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.740542 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-fm2qz"] Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.751620 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-9373-account-create-update-64mfz"] Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.752654 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9373-account-create-update-64mfz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.756999 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.757341 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9373-account-create-update-64mfz"] Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.772220 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkp5b\" (UniqueName: \"kubernetes.io/projected/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-kube-api-access-bkp5b\") pod \"keystone-db-sync-fm2qz\" (UID: \"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1\") " pod="openstack/keystone-db-sync-fm2qz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.772321 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-combined-ca-bundle\") pod \"keystone-db-sync-fm2qz\" (UID: \"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1\") " pod="openstack/keystone-db-sync-fm2qz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.772352 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-config-data\") pod \"keystone-db-sync-fm2qz\" (UID: \"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1\") " pod="openstack/keystone-db-sync-fm2qz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.772395 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dec9942-68fd-4595-826c-9cc4bace11aa-operator-scripts\") pod \"barbican-1814-account-create-update-7l5hz\" (UID: \"6dec9942-68fd-4595-826c-9cc4bace11aa\") " pod="openstack/barbican-1814-account-create-update-7l5hz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.772485 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b6d7\" (UniqueName: \"kubernetes.io/projected/58e52c5d-7e41-4145-abad-945e9cb9d01b-kube-api-access-6b6d7\") pod \"barbican-db-create-4jbjk\" (UID: \"58e52c5d-7e41-4145-abad-945e9cb9d01b\") " pod="openstack/barbican-db-create-4jbjk" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.772555 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1bec2f8e-e989-4408-918b-3b9ad6f01af8-operator-scripts\") pod \"cinder-9373-account-create-update-64mfz\" (UID: \"1bec2f8e-e989-4408-918b-3b9ad6f01af8\") " pod="openstack/cinder-9373-account-create-update-64mfz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.772622 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvtt9\" (UniqueName: \"kubernetes.io/projected/6dec9942-68fd-4595-826c-9cc4bace11aa-kube-api-access-jvtt9\") pod \"barbican-1814-account-create-update-7l5hz\" (UID: \"6dec9942-68fd-4595-826c-9cc4bace11aa\") " pod="openstack/barbican-1814-account-create-update-7l5hz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.772670 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gnv2\" (UniqueName: \"kubernetes.io/projected/1bec2f8e-e989-4408-918b-3b9ad6f01af8-kube-api-access-7gnv2\") pod \"cinder-9373-account-create-update-64mfz\" (UID: \"1bec2f8e-e989-4408-918b-3b9ad6f01af8\") " pod="openstack/cinder-9373-account-create-update-64mfz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.772706 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58e52c5d-7e41-4145-abad-945e9cb9d01b-operator-scripts\") pod \"barbican-db-create-4jbjk\" (UID: \"58e52c5d-7e41-4145-abad-945e9cb9d01b\") " pod="openstack/barbican-db-create-4jbjk" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.773437 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58e52c5d-7e41-4145-abad-945e9cb9d01b-operator-scripts\") pod \"barbican-db-create-4jbjk\" (UID: \"58e52c5d-7e41-4145-abad-945e9cb9d01b\") " pod="openstack/barbican-db-create-4jbjk" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.779949 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6sn86" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.787627 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b6d7\" (UniqueName: \"kubernetes.io/projected/58e52c5d-7e41-4145-abad-945e9cb9d01b-kube-api-access-6b6d7\") pod \"barbican-db-create-4jbjk\" (UID: \"58e52c5d-7e41-4145-abad-945e9cb9d01b\") " pod="openstack/barbican-db-create-4jbjk" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.819862 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-pv8tn"] Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.821254 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pv8tn" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.830315 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pv8tn"] Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.840995 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-f640-account-create-update-vwhg8"] Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.842052 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f640-account-create-update-vwhg8" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.843015 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f640-account-create-update-vwhg8"] Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.868239 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.873990 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvtt9\" (UniqueName: \"kubernetes.io/projected/6dec9942-68fd-4595-826c-9cc4bace11aa-kube-api-access-jvtt9\") pod \"barbican-1814-account-create-update-7l5hz\" (UID: \"6dec9942-68fd-4595-826c-9cc4bace11aa\") " pod="openstack/barbican-1814-account-create-update-7l5hz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.874046 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m2dp\" (UniqueName: \"kubernetes.io/projected/5b93f1ae-e632-4bca-a89e-3ce44858a60e-kube-api-access-9m2dp\") pod \"neutron-db-create-pv8tn\" (UID: \"5b93f1ae-e632-4bca-a89e-3ce44858a60e\") " pod="openstack/neutron-db-create-pv8tn" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.874086 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gnv2\" (UniqueName: \"kubernetes.io/projected/1bec2f8e-e989-4408-918b-3b9ad6f01af8-kube-api-access-7gnv2\") pod \"cinder-9373-account-create-update-64mfz\" (UID: \"1bec2f8e-e989-4408-918b-3b9ad6f01af8\") " pod="openstack/cinder-9373-account-create-update-64mfz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.874138 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b93f1ae-e632-4bca-a89e-3ce44858a60e-operator-scripts\") pod \"neutron-db-create-pv8tn\" (UID: \"5b93f1ae-e632-4bca-a89e-3ce44858a60e\") " pod="openstack/neutron-db-create-pv8tn" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.874160 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a16816d-ed5f-4324-97d5-c30fb6b381bd-operator-scripts\") pod \"neutron-f640-account-create-update-vwhg8\" (UID: \"9a16816d-ed5f-4324-97d5-c30fb6b381bd\") " pod="openstack/neutron-f640-account-create-update-vwhg8" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.874210 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkp5b\" (UniqueName: \"kubernetes.io/projected/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-kube-api-access-bkp5b\") pod \"keystone-db-sync-fm2qz\" (UID: \"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1\") " pod="openstack/keystone-db-sync-fm2qz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.874253 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-combined-ca-bundle\") pod \"keystone-db-sync-fm2qz\" (UID: \"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1\") " pod="openstack/keystone-db-sync-fm2qz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.874272 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-config-data\") pod \"keystone-db-sync-fm2qz\" (UID: \"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1\") " pod="openstack/keystone-db-sync-fm2qz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.874291 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dec9942-68fd-4595-826c-9cc4bace11aa-operator-scripts\") pod \"barbican-1814-account-create-update-7l5hz\" (UID: \"6dec9942-68fd-4595-826c-9cc4bace11aa\") " pod="openstack/barbican-1814-account-create-update-7l5hz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.874329 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1bec2f8e-e989-4408-918b-3b9ad6f01af8-operator-scripts\") pod \"cinder-9373-account-create-update-64mfz\" (UID: \"1bec2f8e-e989-4408-918b-3b9ad6f01af8\") " pod="openstack/cinder-9373-account-create-update-64mfz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.874349 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llmrk\" (UniqueName: \"kubernetes.io/projected/9a16816d-ed5f-4324-97d5-c30fb6b381bd-kube-api-access-llmrk\") pod \"neutron-f640-account-create-update-vwhg8\" (UID: \"9a16816d-ed5f-4324-97d5-c30fb6b381bd\") " pod="openstack/neutron-f640-account-create-update-vwhg8" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.875979 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1bec2f8e-e989-4408-918b-3b9ad6f01af8-operator-scripts\") pod \"cinder-9373-account-create-update-64mfz\" (UID: \"1bec2f8e-e989-4408-918b-3b9ad6f01af8\") " pod="openstack/cinder-9373-account-create-update-64mfz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.876212 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dec9942-68fd-4595-826c-9cc4bace11aa-operator-scripts\") pod \"barbican-1814-account-create-update-7l5hz\" (UID: \"6dec9942-68fd-4595-826c-9cc4bace11aa\") " pod="openstack/barbican-1814-account-create-update-7l5hz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.880672 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-combined-ca-bundle\") pod \"keystone-db-sync-fm2qz\" (UID: \"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1\") " pod="openstack/keystone-db-sync-fm2qz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.891539 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-config-data\") pod \"keystone-db-sync-fm2qz\" (UID: \"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1\") " pod="openstack/keystone-db-sync-fm2qz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.894382 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gnv2\" (UniqueName: \"kubernetes.io/projected/1bec2f8e-e989-4408-918b-3b9ad6f01af8-kube-api-access-7gnv2\") pod \"cinder-9373-account-create-update-64mfz\" (UID: \"1bec2f8e-e989-4408-918b-3b9ad6f01af8\") " pod="openstack/cinder-9373-account-create-update-64mfz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.899150 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvtt9\" (UniqueName: \"kubernetes.io/projected/6dec9942-68fd-4595-826c-9cc4bace11aa-kube-api-access-jvtt9\") pod \"barbican-1814-account-create-update-7l5hz\" (UID: \"6dec9942-68fd-4595-826c-9cc4bace11aa\") " pod="openstack/barbican-1814-account-create-update-7l5hz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.903064 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkp5b\" (UniqueName: \"kubernetes.io/projected/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-kube-api-access-bkp5b\") pod \"keystone-db-sync-fm2qz\" (UID: \"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1\") " pod="openstack/keystone-db-sync-fm2qz" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.972243 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4jbjk" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.975633 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a16816d-ed5f-4324-97d5-c30fb6b381bd-operator-scripts\") pod \"neutron-f640-account-create-update-vwhg8\" (UID: \"9a16816d-ed5f-4324-97d5-c30fb6b381bd\") " pod="openstack/neutron-f640-account-create-update-vwhg8" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.975665 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b93f1ae-e632-4bca-a89e-3ce44858a60e-operator-scripts\") pod \"neutron-db-create-pv8tn\" (UID: \"5b93f1ae-e632-4bca-a89e-3ce44858a60e\") " pod="openstack/neutron-db-create-pv8tn" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.975743 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llmrk\" (UniqueName: \"kubernetes.io/projected/9a16816d-ed5f-4324-97d5-c30fb6b381bd-kube-api-access-llmrk\") pod \"neutron-f640-account-create-update-vwhg8\" (UID: \"9a16816d-ed5f-4324-97d5-c30fb6b381bd\") " pod="openstack/neutron-f640-account-create-update-vwhg8" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.975781 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m2dp\" (UniqueName: \"kubernetes.io/projected/5b93f1ae-e632-4bca-a89e-3ce44858a60e-kube-api-access-9m2dp\") pod \"neutron-db-create-pv8tn\" (UID: \"5b93f1ae-e632-4bca-a89e-3ce44858a60e\") " pod="openstack/neutron-db-create-pv8tn" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.977451 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b93f1ae-e632-4bca-a89e-3ce44858a60e-operator-scripts\") pod \"neutron-db-create-pv8tn\" (UID: \"5b93f1ae-e632-4bca-a89e-3ce44858a60e\") " pod="openstack/neutron-db-create-pv8tn" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.978065 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a16816d-ed5f-4324-97d5-c30fb6b381bd-operator-scripts\") pod \"neutron-f640-account-create-update-vwhg8\" (UID: \"9a16816d-ed5f-4324-97d5-c30fb6b381bd\") " pod="openstack/neutron-f640-account-create-update-vwhg8" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.993981 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m2dp\" (UniqueName: \"kubernetes.io/projected/5b93f1ae-e632-4bca-a89e-3ce44858a60e-kube-api-access-9m2dp\") pod \"neutron-db-create-pv8tn\" (UID: \"5b93f1ae-e632-4bca-a89e-3ce44858a60e\") " pod="openstack/neutron-db-create-pv8tn" Dec 01 15:52:30 crc kubenswrapper[4739]: I1201 15:52:30.995881 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llmrk\" (UniqueName: \"kubernetes.io/projected/9a16816d-ed5f-4324-97d5-c30fb6b381bd-kube-api-access-llmrk\") pod \"neutron-f640-account-create-update-vwhg8\" (UID: \"9a16816d-ed5f-4324-97d5-c30fb6b381bd\") " pod="openstack/neutron-f640-account-create-update-vwhg8" Dec 01 15:52:31 crc kubenswrapper[4739]: I1201 15:52:31.014127 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1814-account-create-update-7l5hz" Dec 01 15:52:31 crc kubenswrapper[4739]: I1201 15:52:31.037090 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-fm2qz" Dec 01 15:52:31 crc kubenswrapper[4739]: I1201 15:52:31.070143 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9373-account-create-update-64mfz" Dec 01 15:52:31 crc kubenswrapper[4739]: I1201 15:52:31.264867 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pv8tn" Dec 01 15:52:31 crc kubenswrapper[4739]: I1201 15:52:31.278196 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f640-account-create-update-vwhg8" Dec 01 15:52:31 crc kubenswrapper[4739]: I1201 15:52:31.281299 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6sn86"] Dec 01 15:52:31 crc kubenswrapper[4739]: I1201 15:52:31.509317 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4jbjk"] Dec 01 15:52:31 crc kubenswrapper[4739]: W1201 15:52:31.520046 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58e52c5d_7e41_4145_abad_945e9cb9d01b.slice/crio-d5dcf62c1e3fdf6fd1d5da58d10740983a2ef2aba005d4abbe40cc4f9ece68a8 WatchSource:0}: Error finding container d5dcf62c1e3fdf6fd1d5da58d10740983a2ef2aba005d4abbe40cc4f9ece68a8: Status 404 returned error can't find the container with id d5dcf62c1e3fdf6fd1d5da58d10740983a2ef2aba005d4abbe40cc4f9ece68a8 Dec 01 15:52:31 crc kubenswrapper[4739]: I1201 15:52:31.610796 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-fm2qz"] Dec 01 15:52:31 crc kubenswrapper[4739]: I1201 15:52:31.670286 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1814-account-create-update-7l5hz"] Dec 01 15:52:31 crc kubenswrapper[4739]: W1201 15:52:31.713621 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dec9942_68fd_4595_826c_9cc4bace11aa.slice/crio-774efdb1bf56833df2f356cd870c7f37c8ae546d9c3a430de7e475630e7f05dd WatchSource:0}: Error finding container 774efdb1bf56833df2f356cd870c7f37c8ae546d9c3a430de7e475630e7f05dd: Status 404 returned error can't find the container with id 774efdb1bf56833df2f356cd870c7f37c8ae546d9c3a430de7e475630e7f05dd Dec 01 15:52:31 crc kubenswrapper[4739]: I1201 15:52:31.731940 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f640-account-create-update-vwhg8"] Dec 01 15:52:31 crc kubenswrapper[4739]: W1201 15:52:31.753328 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a16816d_ed5f_4324_97d5_c30fb6b381bd.slice/crio-1ed299c5338da6c5e76767cd4e009f8caab8980f73ac7c7761ea23804dcce26d WatchSource:0}: Error finding container 1ed299c5338da6c5e76767cd4e009f8caab8980f73ac7c7761ea23804dcce26d: Status 404 returned error can't find the container with id 1ed299c5338da6c5e76767cd4e009f8caab8980f73ac7c7761ea23804dcce26d Dec 01 15:52:31 crc kubenswrapper[4739]: I1201 15:52:31.784288 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9373-account-create-update-64mfz"] Dec 01 15:52:31 crc kubenswrapper[4739]: I1201 15:52:31.865557 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pv8tn"] Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.251915 4739 generic.go:334] "Generic (PLEG): container finished" podID="6dec9942-68fd-4595-826c-9cc4bace11aa" containerID="d47ca4d362eedf2557a6ed7480efc0d503391d899c5e45e9a9b07a5e97240459" exitCode=0 Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.252012 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1814-account-create-update-7l5hz" event={"ID":"6dec9942-68fd-4595-826c-9cc4bace11aa","Type":"ContainerDied","Data":"d47ca4d362eedf2557a6ed7480efc0d503391d899c5e45e9a9b07a5e97240459"} Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.252044 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1814-account-create-update-7l5hz" event={"ID":"6dec9942-68fd-4595-826c-9cc4bace11aa","Type":"ContainerStarted","Data":"774efdb1bf56833df2f356cd870c7f37c8ae546d9c3a430de7e475630e7f05dd"} Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.253645 4739 generic.go:334] "Generic (PLEG): container finished" podID="56bbe10e-0b3e-48c5-a806-cf5f1dec0935" containerID="d2b29e99374d12dfad1e0b93391e46de5fdf34979c0191fb2c48f9d1540bf8cc" exitCode=0 Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.253720 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6sn86" event={"ID":"56bbe10e-0b3e-48c5-a806-cf5f1dec0935","Type":"ContainerDied","Data":"d2b29e99374d12dfad1e0b93391e46de5fdf34979c0191fb2c48f9d1540bf8cc"} Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.253823 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6sn86" event={"ID":"56bbe10e-0b3e-48c5-a806-cf5f1dec0935","Type":"ContainerStarted","Data":"f3da4c4425c14d1a75b4e1b188278be99652abfd59f263d0c40f75d697e256f5"} Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.260783 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9373-account-create-update-64mfz" event={"ID":"1bec2f8e-e989-4408-918b-3b9ad6f01af8","Type":"ContainerStarted","Data":"082718586a23ee88e13b30ddd5d473ead6afadad04225112b281f7207d54fc7d"} Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.260836 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9373-account-create-update-64mfz" event={"ID":"1bec2f8e-e989-4408-918b-3b9ad6f01af8","Type":"ContainerStarted","Data":"aff048b030be2c6e4c482e98ea4236d4a917307df9e2b5ebd8d20bca4b384807"} Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.267791 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-fm2qz" event={"ID":"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1","Type":"ContainerStarted","Data":"bac9da90ca6e35df5260b21f9600c2ce990a1890c7d2355618fd04eb327ba33d"} Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.270836 4739 generic.go:334] "Generic (PLEG): container finished" podID="58e52c5d-7e41-4145-abad-945e9cb9d01b" containerID="7babd0fe406838f7127a2d25735ec535eb730e9836f364ecba8cf00ca14eec41" exitCode=0 Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.271025 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4jbjk" event={"ID":"58e52c5d-7e41-4145-abad-945e9cb9d01b","Type":"ContainerDied","Data":"7babd0fe406838f7127a2d25735ec535eb730e9836f364ecba8cf00ca14eec41"} Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.271060 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4jbjk" event={"ID":"58e52c5d-7e41-4145-abad-945e9cb9d01b","Type":"ContainerStarted","Data":"d5dcf62c1e3fdf6fd1d5da58d10740983a2ef2aba005d4abbe40cc4f9ece68a8"} Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.274494 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f640-account-create-update-vwhg8" event={"ID":"9a16816d-ed5f-4324-97d5-c30fb6b381bd","Type":"ContainerStarted","Data":"49e681dde4cd412b841676f21d70365bdcf6429930e93e74191ebbe9818216f1"} Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.274535 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f640-account-create-update-vwhg8" event={"ID":"9a16816d-ed5f-4324-97d5-c30fb6b381bd","Type":"ContainerStarted","Data":"1ed299c5338da6c5e76767cd4e009f8caab8980f73ac7c7761ea23804dcce26d"} Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.276645 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pv8tn" event={"ID":"5b93f1ae-e632-4bca-a89e-3ce44858a60e","Type":"ContainerStarted","Data":"02d8fb088f94bd4c3c6e53a76b89db02aa2824ba0200335f87ab68cf963628a8"} Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.276678 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pv8tn" event={"ID":"5b93f1ae-e632-4bca-a89e-3ce44858a60e","Type":"ContainerStarted","Data":"37dccd5dc99a79d39fae416e15da3de0a1eff7945ae160e175e2e2cb97edb28b"} Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.288228 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-9373-account-create-update-64mfz" podStartSLOduration=2.288210021 podStartE2EDuration="2.288210021s" podCreationTimestamp="2025-12-01 15:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:52:32.28495247 +0000 UTC m=+1054.110698564" watchObservedRunningTime="2025-12-01 15:52:32.288210021 +0000 UTC m=+1054.113956115" Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.330308 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-pv8tn" podStartSLOduration=2.330293541 podStartE2EDuration="2.330293541s" podCreationTimestamp="2025-12-01 15:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:52:32.324587415 +0000 UTC m=+1054.150333509" watchObservedRunningTime="2025-12-01 15:52:32.330293541 +0000 UTC m=+1054.156039625" Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.345732 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-f640-account-create-update-vwhg8" podStartSLOduration=2.345711067 podStartE2EDuration="2.345711067s" podCreationTimestamp="2025-12-01 15:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:52:32.342079236 +0000 UTC m=+1054.167825330" watchObservedRunningTime="2025-12-01 15:52:32.345711067 +0000 UTC m=+1054.171457161" Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.904773 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.967937 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kcnlh"] Dec 01 15:52:32 crc kubenswrapper[4739]: I1201 15:52:32.968158 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" podUID="2bace9b3-1f79-4e97-8385-580bb4b01ed7" containerName="dnsmasq-dns" containerID="cri-o://c31c7d1d2d67a9db18e15f8088cec205d2b7d85e64134a4ac79d468dd1d0e4ec" gracePeriod=10 Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.293101 4739 generic.go:334] "Generic (PLEG): container finished" podID="5b93f1ae-e632-4bca-a89e-3ce44858a60e" containerID="02d8fb088f94bd4c3c6e53a76b89db02aa2824ba0200335f87ab68cf963628a8" exitCode=0 Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.293230 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pv8tn" event={"ID":"5b93f1ae-e632-4bca-a89e-3ce44858a60e","Type":"ContainerDied","Data":"02d8fb088f94bd4c3c6e53a76b89db02aa2824ba0200335f87ab68cf963628a8"} Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.294694 4739 generic.go:334] "Generic (PLEG): container finished" podID="1bec2f8e-e989-4408-918b-3b9ad6f01af8" containerID="082718586a23ee88e13b30ddd5d473ead6afadad04225112b281f7207d54fc7d" exitCode=0 Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.294749 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9373-account-create-update-64mfz" event={"ID":"1bec2f8e-e989-4408-918b-3b9ad6f01af8","Type":"ContainerDied","Data":"082718586a23ee88e13b30ddd5d473ead6afadad04225112b281f7207d54fc7d"} Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.307016 4739 generic.go:334] "Generic (PLEG): container finished" podID="9a16816d-ed5f-4324-97d5-c30fb6b381bd" containerID="49e681dde4cd412b841676f21d70365bdcf6429930e93e74191ebbe9818216f1" exitCode=0 Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.307148 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f640-account-create-update-vwhg8" event={"ID":"9a16816d-ed5f-4324-97d5-c30fb6b381bd","Type":"ContainerDied","Data":"49e681dde4cd412b841676f21d70365bdcf6429930e93e74191ebbe9818216f1"} Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.317028 4739 generic.go:334] "Generic (PLEG): container finished" podID="2bace9b3-1f79-4e97-8385-580bb4b01ed7" containerID="c31c7d1d2d67a9db18e15f8088cec205d2b7d85e64134a4ac79d468dd1d0e4ec" exitCode=0 Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.317292 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" event={"ID":"2bace9b3-1f79-4e97-8385-580bb4b01ed7","Type":"ContainerDied","Data":"c31c7d1d2d67a9db18e15f8088cec205d2b7d85e64134a4ac79d468dd1d0e4ec"} Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.424125 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.534295 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-ovsdbserver-nb\") pod \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.534481 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvlr7\" (UniqueName: \"kubernetes.io/projected/2bace9b3-1f79-4e97-8385-580bb4b01ed7-kube-api-access-dvlr7\") pod \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.534516 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-dns-svc\") pod \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.534580 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-ovsdbserver-sb\") pod \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.534598 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-config\") pod \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\" (UID: \"2bace9b3-1f79-4e97-8385-580bb4b01ed7\") " Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.554099 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bace9b3-1f79-4e97-8385-580bb4b01ed7-kube-api-access-dvlr7" (OuterVolumeSpecName: "kube-api-access-dvlr7") pod "2bace9b3-1f79-4e97-8385-580bb4b01ed7" (UID: "2bace9b3-1f79-4e97-8385-580bb4b01ed7"). InnerVolumeSpecName "kube-api-access-dvlr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.583755 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2bace9b3-1f79-4e97-8385-580bb4b01ed7" (UID: "2bace9b3-1f79-4e97-8385-580bb4b01ed7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.587839 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2bace9b3-1f79-4e97-8385-580bb4b01ed7" (UID: "2bace9b3-1f79-4e97-8385-580bb4b01ed7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.591082 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-config" (OuterVolumeSpecName: "config") pod "2bace9b3-1f79-4e97-8385-580bb4b01ed7" (UID: "2bace9b3-1f79-4e97-8385-580bb4b01ed7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.601039 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2bace9b3-1f79-4e97-8385-580bb4b01ed7" (UID: "2bace9b3-1f79-4e97-8385-580bb4b01ed7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.638740 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvlr7\" (UniqueName: \"kubernetes.io/projected/2bace9b3-1f79-4e97-8385-580bb4b01ed7-kube-api-access-dvlr7\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.638818 4739 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.638832 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.638841 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.638851 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bace9b3-1f79-4e97-8385-580bb4b01ed7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.737542 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4jbjk" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.826067 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1814-account-create-update-7l5hz" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.834596 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6sn86" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.850327 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvtt9\" (UniqueName: \"kubernetes.io/projected/6dec9942-68fd-4595-826c-9cc4bace11aa-kube-api-access-jvtt9\") pod \"6dec9942-68fd-4595-826c-9cc4bace11aa\" (UID: \"6dec9942-68fd-4595-826c-9cc4bace11aa\") " Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.850467 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dec9942-68fd-4595-826c-9cc4bace11aa-operator-scripts\") pod \"6dec9942-68fd-4595-826c-9cc4bace11aa\" (UID: \"6dec9942-68fd-4595-826c-9cc4bace11aa\") " Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.850618 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b6d7\" (UniqueName: \"kubernetes.io/projected/58e52c5d-7e41-4145-abad-945e9cb9d01b-kube-api-access-6b6d7\") pod \"58e52c5d-7e41-4145-abad-945e9cb9d01b\" (UID: \"58e52c5d-7e41-4145-abad-945e9cb9d01b\") " Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.850741 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58e52c5d-7e41-4145-abad-945e9cb9d01b-operator-scripts\") pod \"58e52c5d-7e41-4145-abad-945e9cb9d01b\" (UID: \"58e52c5d-7e41-4145-abad-945e9cb9d01b\") " Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.851520 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dec9942-68fd-4595-826c-9cc4bace11aa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6dec9942-68fd-4595-826c-9cc4bace11aa" (UID: "6dec9942-68fd-4595-826c-9cc4bace11aa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.852910 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58e52c5d-7e41-4145-abad-945e9cb9d01b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "58e52c5d-7e41-4145-abad-945e9cb9d01b" (UID: "58e52c5d-7e41-4145-abad-945e9cb9d01b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.854328 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58e52c5d-7e41-4145-abad-945e9cb9d01b-kube-api-access-6b6d7" (OuterVolumeSpecName: "kube-api-access-6b6d7") pod "58e52c5d-7e41-4145-abad-945e9cb9d01b" (UID: "58e52c5d-7e41-4145-abad-945e9cb9d01b"). InnerVolumeSpecName "kube-api-access-6b6d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.854996 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dec9942-68fd-4595-826c-9cc4bace11aa-kube-api-access-jvtt9" (OuterVolumeSpecName: "kube-api-access-jvtt9") pod "6dec9942-68fd-4595-826c-9cc4bace11aa" (UID: "6dec9942-68fd-4595-826c-9cc4bace11aa"). InnerVolumeSpecName "kube-api-access-jvtt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.952638 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbzq5\" (UniqueName: \"kubernetes.io/projected/56bbe10e-0b3e-48c5-a806-cf5f1dec0935-kube-api-access-lbzq5\") pod \"56bbe10e-0b3e-48c5-a806-cf5f1dec0935\" (UID: \"56bbe10e-0b3e-48c5-a806-cf5f1dec0935\") " Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.953119 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56bbe10e-0b3e-48c5-a806-cf5f1dec0935-operator-scripts\") pod \"56bbe10e-0b3e-48c5-a806-cf5f1dec0935\" (UID: \"56bbe10e-0b3e-48c5-a806-cf5f1dec0935\") " Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.953628 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvtt9\" (UniqueName: \"kubernetes.io/projected/6dec9942-68fd-4595-826c-9cc4bace11aa-kube-api-access-jvtt9\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.953646 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dec9942-68fd-4595-826c-9cc4bace11aa-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.953663 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b6d7\" (UniqueName: \"kubernetes.io/projected/58e52c5d-7e41-4145-abad-945e9cb9d01b-kube-api-access-6b6d7\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.953653 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56bbe10e-0b3e-48c5-a806-cf5f1dec0935-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "56bbe10e-0b3e-48c5-a806-cf5f1dec0935" (UID: "56bbe10e-0b3e-48c5-a806-cf5f1dec0935"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.953677 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58e52c5d-7e41-4145-abad-945e9cb9d01b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:33 crc kubenswrapper[4739]: I1201 15:52:33.962792 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56bbe10e-0b3e-48c5-a806-cf5f1dec0935-kube-api-access-lbzq5" (OuterVolumeSpecName: "kube-api-access-lbzq5") pod "56bbe10e-0b3e-48c5-a806-cf5f1dec0935" (UID: "56bbe10e-0b3e-48c5-a806-cf5f1dec0935"). InnerVolumeSpecName "kube-api-access-lbzq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.055471 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56bbe10e-0b3e-48c5-a806-cf5f1dec0935-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.055510 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbzq5\" (UniqueName: \"kubernetes.io/projected/56bbe10e-0b3e-48c5-a806-cf5f1dec0935-kube-api-access-lbzq5\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.328865 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" event={"ID":"2bace9b3-1f79-4e97-8385-580bb4b01ed7","Type":"ContainerDied","Data":"0566478c1d41d6201c071d55c408286199f3458485d78c37f212ad03fe6f07c8"} Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.328904 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kcnlh" Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.328943 4739 scope.go:117] "RemoveContainer" containerID="c31c7d1d2d67a9db18e15f8088cec205d2b7d85e64134a4ac79d468dd1d0e4ec" Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.330913 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1814-account-create-update-7l5hz" event={"ID":"6dec9942-68fd-4595-826c-9cc4bace11aa","Type":"ContainerDied","Data":"774efdb1bf56833df2f356cd870c7f37c8ae546d9c3a430de7e475630e7f05dd"} Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.330938 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="774efdb1bf56833df2f356cd870c7f37c8ae546d9c3a430de7e475630e7f05dd" Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.331021 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1814-account-create-update-7l5hz" Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.332859 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6sn86" Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.332855 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6sn86" event={"ID":"56bbe10e-0b3e-48c5-a806-cf5f1dec0935","Type":"ContainerDied","Data":"f3da4c4425c14d1a75b4e1b188278be99652abfd59f263d0c40f75d697e256f5"} Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.332920 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3da4c4425c14d1a75b4e1b188278be99652abfd59f263d0c40f75d697e256f5" Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.336671 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4jbjk" event={"ID":"58e52c5d-7e41-4145-abad-945e9cb9d01b","Type":"ContainerDied","Data":"d5dcf62c1e3fdf6fd1d5da58d10740983a2ef2aba005d4abbe40cc4f9ece68a8"} Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.336696 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5dcf62c1e3fdf6fd1d5da58d10740983a2ef2aba005d4abbe40cc4f9ece68a8" Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.336875 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4jbjk" Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.380524 4739 scope.go:117] "RemoveContainer" containerID="1b699e8807e1dc3e7d9809f3729237f47119de43962ade0d21fab5808e4b1a6a" Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.383534 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kcnlh"] Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.390761 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kcnlh"] Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.517928 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bace9b3-1f79-4e97-8385-580bb4b01ed7" path="/var/lib/kubelet/pods/2bace9b3-1f79-4e97-8385-580bb4b01ed7/volumes" Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.872326 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pv8tn" Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.895354 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9373-account-create-update-64mfz" Dec 01 15:52:34 crc kubenswrapper[4739]: I1201 15:52:34.899962 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f640-account-create-update-vwhg8" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.014185 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m2dp\" (UniqueName: \"kubernetes.io/projected/5b93f1ae-e632-4bca-a89e-3ce44858a60e-kube-api-access-9m2dp\") pod \"5b93f1ae-e632-4bca-a89e-3ce44858a60e\" (UID: \"5b93f1ae-e632-4bca-a89e-3ce44858a60e\") " Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.014341 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1bec2f8e-e989-4408-918b-3b9ad6f01af8-operator-scripts\") pod \"1bec2f8e-e989-4408-918b-3b9ad6f01af8\" (UID: \"1bec2f8e-e989-4408-918b-3b9ad6f01af8\") " Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.014377 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b93f1ae-e632-4bca-a89e-3ce44858a60e-operator-scripts\") pod \"5b93f1ae-e632-4bca-a89e-3ce44858a60e\" (UID: \"5b93f1ae-e632-4bca-a89e-3ce44858a60e\") " Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.014495 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gnv2\" (UniqueName: \"kubernetes.io/projected/1bec2f8e-e989-4408-918b-3b9ad6f01af8-kube-api-access-7gnv2\") pod \"1bec2f8e-e989-4408-918b-3b9ad6f01af8\" (UID: \"1bec2f8e-e989-4408-918b-3b9ad6f01af8\") " Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.014518 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llmrk\" (UniqueName: \"kubernetes.io/projected/9a16816d-ed5f-4324-97d5-c30fb6b381bd-kube-api-access-llmrk\") pod \"9a16816d-ed5f-4324-97d5-c30fb6b381bd\" (UID: \"9a16816d-ed5f-4324-97d5-c30fb6b381bd\") " Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.014553 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a16816d-ed5f-4324-97d5-c30fb6b381bd-operator-scripts\") pod \"9a16816d-ed5f-4324-97d5-c30fb6b381bd\" (UID: \"9a16816d-ed5f-4324-97d5-c30fb6b381bd\") " Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.015300 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bec2f8e-e989-4408-918b-3b9ad6f01af8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1bec2f8e-e989-4408-918b-3b9ad6f01af8" (UID: "1bec2f8e-e989-4408-918b-3b9ad6f01af8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.015380 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a16816d-ed5f-4324-97d5-c30fb6b381bd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9a16816d-ed5f-4324-97d5-c30fb6b381bd" (UID: "9a16816d-ed5f-4324-97d5-c30fb6b381bd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.015411 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b93f1ae-e632-4bca-a89e-3ce44858a60e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5b93f1ae-e632-4bca-a89e-3ce44858a60e" (UID: "5b93f1ae-e632-4bca-a89e-3ce44858a60e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.020934 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bec2f8e-e989-4408-918b-3b9ad6f01af8-kube-api-access-7gnv2" (OuterVolumeSpecName: "kube-api-access-7gnv2") pod "1bec2f8e-e989-4408-918b-3b9ad6f01af8" (UID: "1bec2f8e-e989-4408-918b-3b9ad6f01af8"). InnerVolumeSpecName "kube-api-access-7gnv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.020993 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a16816d-ed5f-4324-97d5-c30fb6b381bd-kube-api-access-llmrk" (OuterVolumeSpecName: "kube-api-access-llmrk") pod "9a16816d-ed5f-4324-97d5-c30fb6b381bd" (UID: "9a16816d-ed5f-4324-97d5-c30fb6b381bd"). InnerVolumeSpecName "kube-api-access-llmrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.021138 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b93f1ae-e632-4bca-a89e-3ce44858a60e-kube-api-access-9m2dp" (OuterVolumeSpecName: "kube-api-access-9m2dp") pod "5b93f1ae-e632-4bca-a89e-3ce44858a60e" (UID: "5b93f1ae-e632-4bca-a89e-3ce44858a60e"). InnerVolumeSpecName "kube-api-access-9m2dp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.116393 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1bec2f8e-e989-4408-918b-3b9ad6f01af8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.116452 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b93f1ae-e632-4bca-a89e-3ce44858a60e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.116463 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gnv2\" (UniqueName: \"kubernetes.io/projected/1bec2f8e-e989-4408-918b-3b9ad6f01af8-kube-api-access-7gnv2\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.116478 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llmrk\" (UniqueName: \"kubernetes.io/projected/9a16816d-ed5f-4324-97d5-c30fb6b381bd-kube-api-access-llmrk\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.116487 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a16816d-ed5f-4324-97d5-c30fb6b381bd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.116496 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m2dp\" (UniqueName: \"kubernetes.io/projected/5b93f1ae-e632-4bca-a89e-3ce44858a60e-kube-api-access-9m2dp\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.348507 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9373-account-create-update-64mfz" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.348515 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9373-account-create-update-64mfz" event={"ID":"1bec2f8e-e989-4408-918b-3b9ad6f01af8","Type":"ContainerDied","Data":"aff048b030be2c6e4c482e98ea4236d4a917307df9e2b5ebd8d20bca4b384807"} Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.348633 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aff048b030be2c6e4c482e98ea4236d4a917307df9e2b5ebd8d20bca4b384807" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.351062 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f640-account-create-update-vwhg8" event={"ID":"9a16816d-ed5f-4324-97d5-c30fb6b381bd","Type":"ContainerDied","Data":"1ed299c5338da6c5e76767cd4e009f8caab8980f73ac7c7761ea23804dcce26d"} Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.351110 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f640-account-create-update-vwhg8" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.351125 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ed299c5338da6c5e76767cd4e009f8caab8980f73ac7c7761ea23804dcce26d" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.354507 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pv8tn" event={"ID":"5b93f1ae-e632-4bca-a89e-3ce44858a60e","Type":"ContainerDied","Data":"37dccd5dc99a79d39fae416e15da3de0a1eff7945ae160e175e2e2cb97edb28b"} Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.354552 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37dccd5dc99a79d39fae416e15da3de0a1eff7945ae160e175e2e2cb97edb28b" Dec 01 15:52:35 crc kubenswrapper[4739]: I1201 15:52:35.354619 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pv8tn" Dec 01 15:52:38 crc kubenswrapper[4739]: I1201 15:52:38.402181 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-fm2qz" event={"ID":"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1","Type":"ContainerStarted","Data":"4acfd69e38cb39550e4d8ff81f3d3ff754b4f0442f2da03146a1e33cdee0c908"} Dec 01 15:52:38 crc kubenswrapper[4739]: I1201 15:52:38.427967 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-fm2qz" podStartSLOduration=2.121576083 podStartE2EDuration="8.427946293s" podCreationTimestamp="2025-12-01 15:52:30 +0000 UTC" firstStartedPulling="2025-12-01 15:52:31.635333122 +0000 UTC m=+1053.461079216" lastFinishedPulling="2025-12-01 15:52:37.941703292 +0000 UTC m=+1059.767449426" observedRunningTime="2025-12-01 15:52:38.427797668 +0000 UTC m=+1060.253543802" watchObservedRunningTime="2025-12-01 15:52:38.427946293 +0000 UTC m=+1060.253692417" Dec 01 15:52:39 crc kubenswrapper[4739]: I1201 15:52:39.621901 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:52:39 crc kubenswrapper[4739]: I1201 15:52:39.622410 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:52:44 crc kubenswrapper[4739]: I1201 15:52:44.462173 4739 generic.go:334] "Generic (PLEG): container finished" podID="3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1" containerID="4acfd69e38cb39550e4d8ff81f3d3ff754b4f0442f2da03146a1e33cdee0c908" exitCode=0 Dec 01 15:52:44 crc kubenswrapper[4739]: I1201 15:52:44.462297 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-fm2qz" event={"ID":"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1","Type":"ContainerDied","Data":"4acfd69e38cb39550e4d8ff81f3d3ff754b4f0442f2da03146a1e33cdee0c908"} Dec 01 15:52:45 crc kubenswrapper[4739]: I1201 15:52:45.871182 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-fm2qz" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.042142 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkp5b\" (UniqueName: \"kubernetes.io/projected/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-kube-api-access-bkp5b\") pod \"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1\" (UID: \"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1\") " Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.042199 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-combined-ca-bundle\") pod \"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1\" (UID: \"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1\") " Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.042344 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-config-data\") pod \"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1\" (UID: \"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1\") " Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.047664 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-kube-api-access-bkp5b" (OuterVolumeSpecName: "kube-api-access-bkp5b") pod "3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1" (UID: "3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1"). InnerVolumeSpecName "kube-api-access-bkp5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.064811 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1" (UID: "3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.083686 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-config-data" (OuterVolumeSpecName: "config-data") pod "3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1" (UID: "3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.145006 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.145058 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkp5b\" (UniqueName: \"kubernetes.io/projected/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-kube-api-access-bkp5b\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.145081 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.486532 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-fm2qz" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.496102 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-fm2qz" event={"ID":"3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1","Type":"ContainerDied","Data":"bac9da90ca6e35df5260b21f9600c2ce990a1890c7d2355618fd04eb327ba33d"} Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.496161 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bac9da90ca6e35df5260b21f9600c2ce990a1890c7d2355618fd04eb327ba33d" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828096 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-hk28c"] Dec 01 15:52:46 crc kubenswrapper[4739]: E1201 15:52:46.828395 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b93f1ae-e632-4bca-a89e-3ce44858a60e" containerName="mariadb-database-create" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828410 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b93f1ae-e632-4bca-a89e-3ce44858a60e" containerName="mariadb-database-create" Dec 01 15:52:46 crc kubenswrapper[4739]: E1201 15:52:46.828441 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56bbe10e-0b3e-48c5-a806-cf5f1dec0935" containerName="mariadb-database-create" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828447 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="56bbe10e-0b3e-48c5-a806-cf5f1dec0935" containerName="mariadb-database-create" Dec 01 15:52:46 crc kubenswrapper[4739]: E1201 15:52:46.828458 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1" containerName="keystone-db-sync" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828464 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1" containerName="keystone-db-sync" Dec 01 15:52:46 crc kubenswrapper[4739]: E1201 15:52:46.828471 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bace9b3-1f79-4e97-8385-580bb4b01ed7" containerName="dnsmasq-dns" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828479 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bace9b3-1f79-4e97-8385-580bb4b01ed7" containerName="dnsmasq-dns" Dec 01 15:52:46 crc kubenswrapper[4739]: E1201 15:52:46.828491 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a16816d-ed5f-4324-97d5-c30fb6b381bd" containerName="mariadb-account-create-update" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828498 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a16816d-ed5f-4324-97d5-c30fb6b381bd" containerName="mariadb-account-create-update" Dec 01 15:52:46 crc kubenswrapper[4739]: E1201 15:52:46.828507 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bec2f8e-e989-4408-918b-3b9ad6f01af8" containerName="mariadb-account-create-update" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828512 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bec2f8e-e989-4408-918b-3b9ad6f01af8" containerName="mariadb-account-create-update" Dec 01 15:52:46 crc kubenswrapper[4739]: E1201 15:52:46.828524 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dec9942-68fd-4595-826c-9cc4bace11aa" containerName="mariadb-account-create-update" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828531 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dec9942-68fd-4595-826c-9cc4bace11aa" containerName="mariadb-account-create-update" Dec 01 15:52:46 crc kubenswrapper[4739]: E1201 15:52:46.828544 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bace9b3-1f79-4e97-8385-580bb4b01ed7" containerName="init" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828551 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bace9b3-1f79-4e97-8385-580bb4b01ed7" containerName="init" Dec 01 15:52:46 crc kubenswrapper[4739]: E1201 15:52:46.828560 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58e52c5d-7e41-4145-abad-945e9cb9d01b" containerName="mariadb-database-create" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828566 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="58e52c5d-7e41-4145-abad-945e9cb9d01b" containerName="mariadb-database-create" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828720 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bec2f8e-e989-4408-918b-3b9ad6f01af8" containerName="mariadb-account-create-update" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828736 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1" containerName="keystone-db-sync" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828742 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a16816d-ed5f-4324-97d5-c30fb6b381bd" containerName="mariadb-account-create-update" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828752 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b93f1ae-e632-4bca-a89e-3ce44858a60e" containerName="mariadb-database-create" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828761 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dec9942-68fd-4595-826c-9cc4bace11aa" containerName="mariadb-account-create-update" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828773 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="56bbe10e-0b3e-48c5-a806-cf5f1dec0935" containerName="mariadb-database-create" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828780 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bace9b3-1f79-4e97-8385-580bb4b01ed7" containerName="dnsmasq-dns" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.828789 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="58e52c5d-7e41-4145-abad-945e9cb9d01b" containerName="mariadb-database-create" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.829280 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.832889 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.833343 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zcbnm" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.833522 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.836470 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.840779 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.847666 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hk28c"] Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.859919 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-z2c5l"] Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.861471 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.878430 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-z2c5l"] Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.958783 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-dns-svc\") pod \"dnsmasq-dns-6546db6db7-z2c5l\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.958828 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-config\") pod \"dnsmasq-dns-6546db6db7-z2c5l\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.958847 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-ovsdbserver-nb\") pod \"dnsmasq-dns-6546db6db7-z2c5l\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.958875 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-credential-keys\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.958976 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cgw2\" (UniqueName: \"kubernetes.io/projected/d144eae4-6ce7-440e-86f0-645a7e943981-kube-api-access-5cgw2\") pod \"dnsmasq-dns-6546db6db7-z2c5l\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.959016 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-fernet-keys\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.959040 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22l4b\" (UniqueName: \"kubernetes.io/projected/c1fca44a-582d-42af-978b-0b38945b260e-kube-api-access-22l4b\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.959116 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-scripts\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.959143 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-ovsdbserver-sb\") pod \"dnsmasq-dns-6546db6db7-z2c5l\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.959172 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-combined-ca-bundle\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:46 crc kubenswrapper[4739]: I1201 15:52:46.959218 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-config-data\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.018609 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.024819 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.026601 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.026788 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.036518 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.050441 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-g9q4b"] Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.051548 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.070529 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-scripts\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.070573 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-ovsdbserver-sb\") pod \"dnsmasq-dns-6546db6db7-z2c5l\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.070999 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-combined-ca-bundle\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.071066 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-config-data\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.071117 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-dns-svc\") pod \"dnsmasq-dns-6546db6db7-z2c5l\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.071141 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-config\") pod \"dnsmasq-dns-6546db6db7-z2c5l\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.071157 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-ovsdbserver-nb\") pod \"dnsmasq-dns-6546db6db7-z2c5l\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.071179 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-credential-keys\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.071219 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cgw2\" (UniqueName: \"kubernetes.io/projected/d144eae4-6ce7-440e-86f0-645a7e943981-kube-api-access-5cgw2\") pod \"dnsmasq-dns-6546db6db7-z2c5l\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.071238 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-fernet-keys\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.071257 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22l4b\" (UniqueName: \"kubernetes.io/projected/c1fca44a-582d-42af-978b-0b38945b260e-kube-api-access-22l4b\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.071453 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-ovsdbserver-sb\") pod \"dnsmasq-dns-6546db6db7-z2c5l\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.072245 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-ovsdbserver-nb\") pod \"dnsmasq-dns-6546db6db7-z2c5l\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.072908 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.075627 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-scripts\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.076615 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5dk2j" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.076781 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.081277 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-config-data\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.100131 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-combined-ca-bundle\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.106365 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22l4b\" (UniqueName: \"kubernetes.io/projected/c1fca44a-582d-42af-978b-0b38945b260e-kube-api-access-22l4b\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.111440 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-dns-svc\") pod \"dnsmasq-dns-6546db6db7-z2c5l\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.113084 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-credential-keys\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.120019 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-fernet-keys\") pod \"keystone-bootstrap-hk28c\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.123978 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-config\") pod \"dnsmasq-dns-6546db6db7-z2c5l\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.127457 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-g9q4b"] Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.146156 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cgw2\" (UniqueName: \"kubernetes.io/projected/d144eae4-6ce7-440e-86f0-645a7e943981-kube-api-access-5cgw2\") pod \"dnsmasq-dns-6546db6db7-z2c5l\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.146304 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-28jb8"] Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.153550 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-28jb8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.154267 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-kfsn8"] Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.155731 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-kfsn8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.156471 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.156982 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-r8k2p" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.156994 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.167892 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.168157 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.168307 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-g66cf" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.187473 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-etc-machine-id\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.187538 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.187558 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbp5v\" (UniqueName: \"kubernetes.io/projected/f38003e3-7781-4602-89e0-0b89e7cf093b-kube-api-access-vbp5v\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.188376 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.190114 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-combined-ca-bundle\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.190144 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-scripts\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.190193 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-config-data\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.190216 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8s7v\" (UniqueName: \"kubernetes.io/projected/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-kube-api-access-v8s7v\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.190317 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f38003e3-7781-4602-89e0-0b89e7cf093b-run-httpd\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.190336 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-db-sync-config-data\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.190371 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-scripts\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.190399 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.190438 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f38003e3-7781-4602-89e0-0b89e7cf093b-log-httpd\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.190495 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-config-data\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.204286 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-28jb8"] Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.210790 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-kfsn8"] Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.291798 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-scripts\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292061 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292090 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f38003e3-7781-4602-89e0-0b89e7cf093b-log-httpd\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292118 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k844l\" (UniqueName: \"kubernetes.io/projected/c2999b95-023b-479c-a18c-0646e3e414d2-kube-api-access-k844l\") pod \"neutron-db-sync-28jb8\" (UID: \"c2999b95-023b-479c-a18c-0646e3e414d2\") " pod="openstack/neutron-db-sync-28jb8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292157 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2999b95-023b-479c-a18c-0646e3e414d2-combined-ca-bundle\") pod \"neutron-db-sync-28jb8\" (UID: \"c2999b95-023b-479c-a18c-0646e3e414d2\") " pod="openstack/neutron-db-sync-28jb8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292179 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-config-data\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292202 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-etc-machine-id\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292220 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0cd4721c-4dfb-4219-9c69-c46299409e2a-db-sync-config-data\") pod \"barbican-db-sync-kfsn8\" (UID: \"0cd4721c-4dfb-4219-9c69-c46299409e2a\") " pod="openstack/barbican-db-sync-kfsn8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292244 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292265 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbp5v\" (UniqueName: \"kubernetes.io/projected/f38003e3-7781-4602-89e0-0b89e7cf093b-kube-api-access-vbp5v\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292290 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jk9x\" (UniqueName: \"kubernetes.io/projected/0cd4721c-4dfb-4219-9c69-c46299409e2a-kube-api-access-4jk9x\") pod \"barbican-db-sync-kfsn8\" (UID: \"0cd4721c-4dfb-4219-9c69-c46299409e2a\") " pod="openstack/barbican-db-sync-kfsn8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292314 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-combined-ca-bundle\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292330 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-scripts\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292349 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-config-data\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292364 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8s7v\" (UniqueName: \"kubernetes.io/projected/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-kube-api-access-v8s7v\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292397 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cd4721c-4dfb-4219-9c69-c46299409e2a-combined-ca-bundle\") pod \"barbican-db-sync-kfsn8\" (UID: \"0cd4721c-4dfb-4219-9c69-c46299409e2a\") " pod="openstack/barbican-db-sync-kfsn8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292454 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f38003e3-7781-4602-89e0-0b89e7cf093b-run-httpd\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292473 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c2999b95-023b-479c-a18c-0646e3e414d2-config\") pod \"neutron-db-sync-28jb8\" (UID: \"c2999b95-023b-479c-a18c-0646e3e414d2\") " pod="openstack/neutron-db-sync-28jb8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.292489 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-db-sync-config-data\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.295015 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f38003e3-7781-4602-89e0-0b89e7cf093b-log-httpd\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.298652 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-etc-machine-id\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.299302 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f38003e3-7781-4602-89e0-0b89e7cf093b-run-httpd\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.299764 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.301010 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-z2c5l"] Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.306869 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-config-data\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.312387 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-scripts\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.319581 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-scripts\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.319786 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.320314 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-combined-ca-bundle\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.332792 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-config-data\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.335226 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-db-sync-config-data\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.342143 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbp5v\" (UniqueName: \"kubernetes.io/projected/f38003e3-7781-4602-89e0-0b89e7cf093b-kube-api-access-vbp5v\") pod \"ceilometer-0\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.344843 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-2lnl5"] Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.345943 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.362250 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.362436 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-gxs8g" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.362590 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.362731 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8s7v\" (UniqueName: \"kubernetes.io/projected/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-kube-api-access-v8s7v\") pod \"cinder-db-sync-g9q4b\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.363133 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-2lnl5"] Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.374493 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-gz97r"] Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.375797 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.388877 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-gz97r"] Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.393398 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cd4721c-4dfb-4219-9c69-c46299409e2a-combined-ca-bundle\") pod \"barbican-db-sync-kfsn8\" (UID: \"0cd4721c-4dfb-4219-9c69-c46299409e2a\") " pod="openstack/barbican-db-sync-kfsn8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.393458 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c2999b95-023b-479c-a18c-0646e3e414d2-config\") pod \"neutron-db-sync-28jb8\" (UID: \"c2999b95-023b-479c-a18c-0646e3e414d2\") " pod="openstack/neutron-db-sync-28jb8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.394323 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k844l\" (UniqueName: \"kubernetes.io/projected/c2999b95-023b-479c-a18c-0646e3e414d2-kube-api-access-k844l\") pod \"neutron-db-sync-28jb8\" (UID: \"c2999b95-023b-479c-a18c-0646e3e414d2\") " pod="openstack/neutron-db-sync-28jb8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.394392 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2999b95-023b-479c-a18c-0646e3e414d2-combined-ca-bundle\") pod \"neutron-db-sync-28jb8\" (UID: \"c2999b95-023b-479c-a18c-0646e3e414d2\") " pod="openstack/neutron-db-sync-28jb8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.394454 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0cd4721c-4dfb-4219-9c69-c46299409e2a-db-sync-config-data\") pod \"barbican-db-sync-kfsn8\" (UID: \"0cd4721c-4dfb-4219-9c69-c46299409e2a\") " pod="openstack/barbican-db-sync-kfsn8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.394509 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jk9x\" (UniqueName: \"kubernetes.io/projected/0cd4721c-4dfb-4219-9c69-c46299409e2a-kube-api-access-4jk9x\") pod \"barbican-db-sync-kfsn8\" (UID: \"0cd4721c-4dfb-4219-9c69-c46299409e2a\") " pod="openstack/barbican-db-sync-kfsn8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.400072 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2999b95-023b-479c-a18c-0646e3e414d2-combined-ca-bundle\") pod \"neutron-db-sync-28jb8\" (UID: \"c2999b95-023b-479c-a18c-0646e3e414d2\") " pod="openstack/neutron-db-sync-28jb8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.405862 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0cd4721c-4dfb-4219-9c69-c46299409e2a-db-sync-config-data\") pod \"barbican-db-sync-kfsn8\" (UID: \"0cd4721c-4dfb-4219-9c69-c46299409e2a\") " pod="openstack/barbican-db-sync-kfsn8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.406044 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c2999b95-023b-479c-a18c-0646e3e414d2-config\") pod \"neutron-db-sync-28jb8\" (UID: \"c2999b95-023b-479c-a18c-0646e3e414d2\") " pod="openstack/neutron-db-sync-28jb8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.408457 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cd4721c-4dfb-4219-9c69-c46299409e2a-combined-ca-bundle\") pod \"barbican-db-sync-kfsn8\" (UID: \"0cd4721c-4dfb-4219-9c69-c46299409e2a\") " pod="openstack/barbican-db-sync-kfsn8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.459033 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jk9x\" (UniqueName: \"kubernetes.io/projected/0cd4721c-4dfb-4219-9c69-c46299409e2a-kube-api-access-4jk9x\") pod \"barbican-db-sync-kfsn8\" (UID: \"0cd4721c-4dfb-4219-9c69-c46299409e2a\") " pod="openstack/barbican-db-sync-kfsn8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.483019 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k844l\" (UniqueName: \"kubernetes.io/projected/c2999b95-023b-479c-a18c-0646e3e414d2-kube-api-access-k844l\") pod \"neutron-db-sync-28jb8\" (UID: \"c2999b95-023b-479c-a18c-0646e3e414d2\") " pod="openstack/neutron-db-sync-28jb8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.496288 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7b5q\" (UniqueName: \"kubernetes.io/projected/6768159c-c437-4457-8d53-2032c1a9662e-kube-api-access-m7b5q\") pod \"dnsmasq-dns-7987f74bbc-gz97r\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.496328 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-combined-ca-bundle\") pod \"placement-db-sync-2lnl5\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.496352 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-gz97r\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.496390 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-gz97r\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.496438 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmdhv\" (UniqueName: \"kubernetes.io/projected/baaebe35-daa8-449a-85fc-069056a1e9bc-kube-api-access-xmdhv\") pod \"placement-db-sync-2lnl5\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.496475 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-scripts\") pod \"placement-db-sync-2lnl5\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.496511 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-config-data\") pod \"placement-db-sync-2lnl5\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.496533 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/baaebe35-daa8-449a-85fc-069056a1e9bc-logs\") pod \"placement-db-sync-2lnl5\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.496564 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-config\") pod \"dnsmasq-dns-7987f74bbc-gz97r\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.496587 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-gz97r\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.550589 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.589787 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.597912 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-config-data\") pod \"placement-db-sync-2lnl5\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.597955 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/baaebe35-daa8-449a-85fc-069056a1e9bc-logs\") pod \"placement-db-sync-2lnl5\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.598021 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-config\") pod \"dnsmasq-dns-7987f74bbc-gz97r\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.598052 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-gz97r\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.598067 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7b5q\" (UniqueName: \"kubernetes.io/projected/6768159c-c437-4457-8d53-2032c1a9662e-kube-api-access-m7b5q\") pod \"dnsmasq-dns-7987f74bbc-gz97r\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.598083 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-combined-ca-bundle\") pod \"placement-db-sync-2lnl5\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.598098 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-gz97r\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.598147 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-gz97r\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.598202 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmdhv\" (UniqueName: \"kubernetes.io/projected/baaebe35-daa8-449a-85fc-069056a1e9bc-kube-api-access-xmdhv\") pod \"placement-db-sync-2lnl5\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.598480 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-scripts\") pod \"placement-db-sync-2lnl5\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.601083 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-ovsdbserver-nb\") pod \"dnsmasq-dns-7987f74bbc-gz97r\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.602038 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-ovsdbserver-sb\") pod \"dnsmasq-dns-7987f74bbc-gz97r\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.602710 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/baaebe35-daa8-449a-85fc-069056a1e9bc-logs\") pod \"placement-db-sync-2lnl5\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.604923 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-combined-ca-bundle\") pod \"placement-db-sync-2lnl5\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.605205 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-dns-svc\") pod \"dnsmasq-dns-7987f74bbc-gz97r\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.605288 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-config-data\") pod \"placement-db-sync-2lnl5\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.608780 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-scripts\") pod \"placement-db-sync-2lnl5\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.615931 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-config\") pod \"dnsmasq-dns-7987f74bbc-gz97r\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.635149 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7b5q\" (UniqueName: \"kubernetes.io/projected/6768159c-c437-4457-8d53-2032c1a9662e-kube-api-access-m7b5q\") pod \"dnsmasq-dns-7987f74bbc-gz97r\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.637169 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmdhv\" (UniqueName: \"kubernetes.io/projected/baaebe35-daa8-449a-85fc-069056a1e9bc-kube-api-access-xmdhv\") pod \"placement-db-sync-2lnl5\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.666834 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-28jb8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.687922 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-kfsn8" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.709818 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2lnl5" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.742334 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:47 crc kubenswrapper[4739]: I1201 15:52:47.965156 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hk28c"] Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.058514 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-z2c5l"] Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.202191 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.294934 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-28jb8"] Dec 01 15:52:48 crc kubenswrapper[4739]: W1201 15:52:48.300128 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2999b95_023b_479c_a18c_0646e3e414d2.slice/crio-5fb1f452d7536f7c64e149f9ab197834eef24d2d7530109abac42eda5ae5e594 WatchSource:0}: Error finding container 5fb1f452d7536f7c64e149f9ab197834eef24d2d7530109abac42eda5ae5e594: Status 404 returned error can't find the container with id 5fb1f452d7536f7c64e149f9ab197834eef24d2d7530109abac42eda5ae5e594 Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.310498 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-g9q4b"] Dec 01 15:52:48 crc kubenswrapper[4739]: W1201 15:52:48.383590 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cd4721c_4dfb_4219_9c69_c46299409e2a.slice/crio-989c4525a84b9c9ad7400f582965fd696508e84251081679096e077f8819060d WatchSource:0}: Error finding container 989c4525a84b9c9ad7400f582965fd696508e84251081679096e077f8819060d: Status 404 returned error can't find the container with id 989c4525a84b9c9ad7400f582965fd696508e84251081679096e077f8819060d Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.385360 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-kfsn8"] Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.411467 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-2lnl5"] Dec 01 15:52:48 crc kubenswrapper[4739]: W1201 15:52:48.426747 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbaaebe35_daa8_449a_85fc_069056a1e9bc.slice/crio-eb213b2c2ce901c1c6e5bbd2168ab8adda9b175bd0ef2b1f83857c5ca9113468 WatchSource:0}: Error finding container eb213b2c2ce901c1c6e5bbd2168ab8adda9b175bd0ef2b1f83857c5ca9113468: Status 404 returned error can't find the container with id eb213b2c2ce901c1c6e5bbd2168ab8adda9b175bd0ef2b1f83857c5ca9113468 Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.546591 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-kfsn8" event={"ID":"0cd4721c-4dfb-4219-9c69-c46299409e2a","Type":"ContainerStarted","Data":"989c4525a84b9c9ad7400f582965fd696508e84251081679096e077f8819060d"} Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.554957 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hk28c" event={"ID":"c1fca44a-582d-42af-978b-0b38945b260e","Type":"ContainerStarted","Data":"9665bb75e988ad1593824068baacfbd407afe31d88d02b82b9fbe975a9a2329d"} Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.555001 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hk28c" event={"ID":"c1fca44a-582d-42af-978b-0b38945b260e","Type":"ContainerStarted","Data":"33e8d209fcfbc209edf3e91b6281076a831d130498554c863a1d7c9a2c5519a8"} Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.564495 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-28jb8" event={"ID":"c2999b95-023b-479c-a18c-0646e3e414d2","Type":"ContainerStarted","Data":"4cc201fe0207deb7ae939f6957d1ab463f42af0ddc04536d6010041aaecbb571"} Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.564556 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-28jb8" event={"ID":"c2999b95-023b-479c-a18c-0646e3e414d2","Type":"ContainerStarted","Data":"5fb1f452d7536f7c64e149f9ab197834eef24d2d7530109abac42eda5ae5e594"} Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.566442 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-g9q4b" event={"ID":"588fd2a1-4b85-4af6-8b48-eff7fdf729ba","Type":"ContainerStarted","Data":"5b161810ddf4e767c15b392ce731a74cf2e75d8b50bd1d5a39efec9d91023ad0"} Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.566984 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-gz97r"] Dec 01 15:52:48 crc kubenswrapper[4739]: W1201 15:52:48.567515 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6768159c_c437_4457_8d53_2032c1a9662e.slice/crio-7c04d2d730534cfa355a211cc1cefd2d94d2e7230e93b3b3357853c64d9aa3a5 WatchSource:0}: Error finding container 7c04d2d730534cfa355a211cc1cefd2d94d2e7230e93b3b3357853c64d9aa3a5: Status 404 returned error can't find the container with id 7c04d2d730534cfa355a211cc1cefd2d94d2e7230e93b3b3357853c64d9aa3a5 Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.568616 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2lnl5" event={"ID":"baaebe35-daa8-449a-85fc-069056a1e9bc","Type":"ContainerStarted","Data":"eb213b2c2ce901c1c6e5bbd2168ab8adda9b175bd0ef2b1f83857c5ca9113468"} Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.570454 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f38003e3-7781-4602-89e0-0b89e7cf093b","Type":"ContainerStarted","Data":"7841d2ec29d1b4929d6f3fa922532bc9481553b1f8303470a319c4e4ee4f752f"} Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.575338 4739 generic.go:334] "Generic (PLEG): container finished" podID="d144eae4-6ce7-440e-86f0-645a7e943981" containerID="794e3c05df2cf7318ecca17479e1e4c5d61b64e395e56bc8fbb2bdcaa20d2298" exitCode=0 Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.575377 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" event={"ID":"d144eae4-6ce7-440e-86f0-645a7e943981","Type":"ContainerDied","Data":"794e3c05df2cf7318ecca17479e1e4c5d61b64e395e56bc8fbb2bdcaa20d2298"} Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.575399 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" event={"ID":"d144eae4-6ce7-440e-86f0-645a7e943981","Type":"ContainerStarted","Data":"ae20fec2eb8eeb4717136118730273ab64d8a94a62ccd0a0fa3f7a02f3df1e4a"} Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.654094 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-28jb8" podStartSLOduration=1.654074734 podStartE2EDuration="1.654074734s" podCreationTimestamp="2025-12-01 15:52:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:52:48.643338052 +0000 UTC m=+1070.469084146" watchObservedRunningTime="2025-12-01 15:52:48.654074734 +0000 UTC m=+1070.479820828" Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.705118 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-hk28c" podStartSLOduration=2.70509909 podStartE2EDuration="2.70509909s" podCreationTimestamp="2025-12-01 15:52:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:52:48.698161585 +0000 UTC m=+1070.523907679" watchObservedRunningTime="2025-12-01 15:52:48.70509909 +0000 UTC m=+1070.530845184" Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.866067 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.931042 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-ovsdbserver-sb\") pod \"d144eae4-6ce7-440e-86f0-645a7e943981\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.931094 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-config\") pod \"d144eae4-6ce7-440e-86f0-645a7e943981\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.931123 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cgw2\" (UniqueName: \"kubernetes.io/projected/d144eae4-6ce7-440e-86f0-645a7e943981-kube-api-access-5cgw2\") pod \"d144eae4-6ce7-440e-86f0-645a7e943981\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.931157 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-dns-svc\") pod \"d144eae4-6ce7-440e-86f0-645a7e943981\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.931179 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-ovsdbserver-nb\") pod \"d144eae4-6ce7-440e-86f0-645a7e943981\" (UID: \"d144eae4-6ce7-440e-86f0-645a7e943981\") " Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.963148 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d144eae4-6ce7-440e-86f0-645a7e943981-kube-api-access-5cgw2" (OuterVolumeSpecName: "kube-api-access-5cgw2") pod "d144eae4-6ce7-440e-86f0-645a7e943981" (UID: "d144eae4-6ce7-440e-86f0-645a7e943981"). InnerVolumeSpecName "kube-api-access-5cgw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.968703 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d144eae4-6ce7-440e-86f0-645a7e943981" (UID: "d144eae4-6ce7-440e-86f0-645a7e943981"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.973455 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-config" (OuterVolumeSpecName: "config") pod "d144eae4-6ce7-440e-86f0-645a7e943981" (UID: "d144eae4-6ce7-440e-86f0-645a7e943981"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:48 crc kubenswrapper[4739]: I1201 15:52:48.994060 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d144eae4-6ce7-440e-86f0-645a7e943981" (UID: "d144eae4-6ce7-440e-86f0-645a7e943981"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:49 crc kubenswrapper[4739]: I1201 15:52:49.011375 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:52:49 crc kubenswrapper[4739]: I1201 15:52:49.027658 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d144eae4-6ce7-440e-86f0-645a7e943981" (UID: "d144eae4-6ce7-440e-86f0-645a7e943981"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:52:49 crc kubenswrapper[4739]: I1201 15:52:49.032069 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:49 crc kubenswrapper[4739]: I1201 15:52:49.032095 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:49 crc kubenswrapper[4739]: I1201 15:52:49.032104 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cgw2\" (UniqueName: \"kubernetes.io/projected/d144eae4-6ce7-440e-86f0-645a7e943981-kube-api-access-5cgw2\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:49 crc kubenswrapper[4739]: I1201 15:52:49.032115 4739 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:49 crc kubenswrapper[4739]: I1201 15:52:49.032123 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d144eae4-6ce7-440e-86f0-645a7e943981-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 15:52:49 crc kubenswrapper[4739]: I1201 15:52:49.589853 4739 generic.go:334] "Generic (PLEG): container finished" podID="6768159c-c437-4457-8d53-2032c1a9662e" containerID="756a36be2e7180142ff4d2b3fc729d651f365c06a27ad8815c7af4f15cd1b634" exitCode=0 Dec 01 15:52:49 crc kubenswrapper[4739]: I1201 15:52:49.590187 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" event={"ID":"6768159c-c437-4457-8d53-2032c1a9662e","Type":"ContainerDied","Data":"756a36be2e7180142ff4d2b3fc729d651f365c06a27ad8815c7af4f15cd1b634"} Dec 01 15:52:49 crc kubenswrapper[4739]: I1201 15:52:49.590213 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" event={"ID":"6768159c-c437-4457-8d53-2032c1a9662e","Type":"ContainerStarted","Data":"7c04d2d730534cfa355a211cc1cefd2d94d2e7230e93b3b3357853c64d9aa3a5"} Dec 01 15:52:49 crc kubenswrapper[4739]: I1201 15:52:49.595963 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" event={"ID":"d144eae4-6ce7-440e-86f0-645a7e943981","Type":"ContainerDied","Data":"ae20fec2eb8eeb4717136118730273ab64d8a94a62ccd0a0fa3f7a02f3df1e4a"} Dec 01 15:52:49 crc kubenswrapper[4739]: I1201 15:52:49.596030 4739 scope.go:117] "RemoveContainer" containerID="794e3c05df2cf7318ecca17479e1e4c5d61b64e395e56bc8fbb2bdcaa20d2298" Dec 01 15:52:49 crc kubenswrapper[4739]: I1201 15:52:49.596315 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6546db6db7-z2c5l" Dec 01 15:52:50 crc kubenswrapper[4739]: I1201 15:52:50.232663 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-z2c5l"] Dec 01 15:52:50 crc kubenswrapper[4739]: I1201 15:52:50.240506 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6546db6db7-z2c5l"] Dec 01 15:52:50 crc kubenswrapper[4739]: I1201 15:52:50.491453 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d144eae4-6ce7-440e-86f0-645a7e943981" path="/var/lib/kubelet/pods/d144eae4-6ce7-440e-86f0-645a7e943981/volumes" Dec 01 15:52:50 crc kubenswrapper[4739]: I1201 15:52:50.611484 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" event={"ID":"6768159c-c437-4457-8d53-2032c1a9662e","Type":"ContainerStarted","Data":"7ff5bc3a6f896fa2fcbf20d72dae92a9285a7ac2c8e6dccfa428ffd20842dd80"} Dec 01 15:52:50 crc kubenswrapper[4739]: I1201 15:52:50.611634 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:50 crc kubenswrapper[4739]: I1201 15:52:50.635119 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" podStartSLOduration=3.635101613 podStartE2EDuration="3.635101613s" podCreationTimestamp="2025-12-01 15:52:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:52:50.633486803 +0000 UTC m=+1072.459232897" watchObservedRunningTime="2025-12-01 15:52:50.635101613 +0000 UTC m=+1072.460847707" Dec 01 15:52:51 crc kubenswrapper[4739]: I1201 15:52:51.619875 4739 generic.go:334] "Generic (PLEG): container finished" podID="c1fca44a-582d-42af-978b-0b38945b260e" containerID="9665bb75e988ad1593824068baacfbd407afe31d88d02b82b9fbe975a9a2329d" exitCode=0 Dec 01 15:52:51 crc kubenswrapper[4739]: I1201 15:52:51.620053 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hk28c" event={"ID":"c1fca44a-582d-42af-978b-0b38945b260e","Type":"ContainerDied","Data":"9665bb75e988ad1593824068baacfbd407afe31d88d02b82b9fbe975a9a2329d"} Dec 01 15:52:57 crc kubenswrapper[4739]: I1201 15:52:57.744587 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:52:57 crc kubenswrapper[4739]: I1201 15:52:57.811662 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-nqkgc"] Dec 01 15:52:57 crc kubenswrapper[4739]: I1201 15:52:57.827665 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" podUID="9ffc2832-c824-470f-8cfa-9ad4ea814125" containerName="dnsmasq-dns" containerID="cri-o://9386ffe9422d13a240a8f5e843765882e844a274968e8d74e834de01f6d2566c" gracePeriod=10 Dec 01 15:52:57 crc kubenswrapper[4739]: I1201 15:52:57.903857 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" podUID="9ffc2832-c824-470f-8cfa-9ad4ea814125" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: connect: connection refused" Dec 01 15:52:58 crc kubenswrapper[4739]: I1201 15:52:58.697394 4739 generic.go:334] "Generic (PLEG): container finished" podID="9ffc2832-c824-470f-8cfa-9ad4ea814125" containerID="9386ffe9422d13a240a8f5e843765882e844a274968e8d74e834de01f6d2566c" exitCode=0 Dec 01 15:52:58 crc kubenswrapper[4739]: I1201 15:52:58.697454 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" event={"ID":"9ffc2832-c824-470f-8cfa-9ad4ea814125","Type":"ContainerDied","Data":"9386ffe9422d13a240a8f5e843765882e844a274968e8d74e834de01f6d2566c"} Dec 01 15:53:02 crc kubenswrapper[4739]: I1201 15:53:02.902970 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" podUID="9ffc2832-c824-470f-8cfa-9ad4ea814125" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: connect: connection refused" Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.792277 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hk28c" event={"ID":"c1fca44a-582d-42af-978b-0b38945b260e","Type":"ContainerDied","Data":"33e8d209fcfbc209edf3e91b6281076a831d130498554c863a1d7c9a2c5519a8"} Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.792858 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33e8d209fcfbc209edf3e91b6281076a831d130498554c863a1d7c9a2c5519a8" Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.795548 4739 generic.go:334] "Generic (PLEG): container finished" podID="c2999b95-023b-479c-a18c-0646e3e414d2" containerID="4cc201fe0207deb7ae939f6957d1ab463f42af0ddc04536d6010041aaecbb571" exitCode=0 Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.795585 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-28jb8" event={"ID":"c2999b95-023b-479c-a18c-0646e3e414d2","Type":"ContainerDied","Data":"4cc201fe0207deb7ae939f6957d1ab463f42af0ddc04536d6010041aaecbb571"} Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.818935 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.868034 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-credential-keys\") pod \"c1fca44a-582d-42af-978b-0b38945b260e\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.868069 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-fernet-keys\") pod \"c1fca44a-582d-42af-978b-0b38945b260e\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.868114 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-combined-ca-bundle\") pod \"c1fca44a-582d-42af-978b-0b38945b260e\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.868151 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22l4b\" (UniqueName: \"kubernetes.io/projected/c1fca44a-582d-42af-978b-0b38945b260e-kube-api-access-22l4b\") pod \"c1fca44a-582d-42af-978b-0b38945b260e\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.869092 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-scripts\") pod \"c1fca44a-582d-42af-978b-0b38945b260e\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.869137 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-config-data\") pod \"c1fca44a-582d-42af-978b-0b38945b260e\" (UID: \"c1fca44a-582d-42af-978b-0b38945b260e\") " Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.876303 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c1fca44a-582d-42af-978b-0b38945b260e" (UID: "c1fca44a-582d-42af-978b-0b38945b260e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.877503 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-scripts" (OuterVolumeSpecName: "scripts") pod "c1fca44a-582d-42af-978b-0b38945b260e" (UID: "c1fca44a-582d-42af-978b-0b38945b260e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.877986 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1fca44a-582d-42af-978b-0b38945b260e-kube-api-access-22l4b" (OuterVolumeSpecName: "kube-api-access-22l4b") pod "c1fca44a-582d-42af-978b-0b38945b260e" (UID: "c1fca44a-582d-42af-978b-0b38945b260e"). InnerVolumeSpecName "kube-api-access-22l4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.879456 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c1fca44a-582d-42af-978b-0b38945b260e" (UID: "c1fca44a-582d-42af-978b-0b38945b260e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.896728 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-config-data" (OuterVolumeSpecName: "config-data") pod "c1fca44a-582d-42af-978b-0b38945b260e" (UID: "c1fca44a-582d-42af-978b-0b38945b260e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.904365 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1fca44a-582d-42af-978b-0b38945b260e" (UID: "c1fca44a-582d-42af-978b-0b38945b260e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.971162 4739 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.971204 4739 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.971217 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.971232 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22l4b\" (UniqueName: \"kubernetes.io/projected/c1fca44a-582d-42af-978b-0b38945b260e-kube-api-access-22l4b\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.971248 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:07 crc kubenswrapper[4739]: I1201 15:53:07.971260 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1fca44a-582d-42af-978b-0b38945b260e-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:08 crc kubenswrapper[4739]: E1201 15:53:08.325196 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 01 15:53:08 crc kubenswrapper[4739]: E1201 15:53:08.325598 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4jk9x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-kfsn8_openstack(0cd4721c-4dfb-4219-9c69-c46299409e2a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 15:53:08 crc kubenswrapper[4739]: E1201 15:53:08.327581 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-kfsn8" podUID="0cd4721c-4dfb-4219-9c69-c46299409e2a" Dec 01 15:53:08 crc kubenswrapper[4739]: I1201 15:53:08.805436 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hk28c" Dec 01 15:53:08 crc kubenswrapper[4739]: E1201 15:53:08.808519 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-kfsn8" podUID="0cd4721c-4dfb-4219-9c69-c46299409e2a" Dec 01 15:53:08 crc kubenswrapper[4739]: I1201 15:53:08.991763 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-hk28c"] Dec 01 15:53:08 crc kubenswrapper[4739]: I1201 15:53:08.996747 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-hk28c"] Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.082335 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-ztf2h"] Dec 01 15:53:09 crc kubenswrapper[4739]: E1201 15:53:09.082987 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d144eae4-6ce7-440e-86f0-645a7e943981" containerName="init" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.083000 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="d144eae4-6ce7-440e-86f0-645a7e943981" containerName="init" Dec 01 15:53:09 crc kubenswrapper[4739]: E1201 15:53:09.083017 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1fca44a-582d-42af-978b-0b38945b260e" containerName="keystone-bootstrap" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.083024 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1fca44a-582d-42af-978b-0b38945b260e" containerName="keystone-bootstrap" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.083162 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="d144eae4-6ce7-440e-86f0-645a7e943981" containerName="init" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.083176 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1fca44a-582d-42af-978b-0b38945b260e" containerName="keystone-bootstrap" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.083675 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.088978 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.089066 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.089128 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.089128 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.089491 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zcbnm" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.092021 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-config-data\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.092115 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-combined-ca-bundle\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.092160 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hft4g\" (UniqueName: \"kubernetes.io/projected/87075836-5cb8-4014-b7f9-04c16c3e9e49-kube-api-access-hft4g\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.092197 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-fernet-keys\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.092299 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-credential-keys\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.092355 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-scripts\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.111092 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ztf2h"] Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.195151 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-credential-keys\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.195218 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-scripts\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.195247 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-config-data\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.195275 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-combined-ca-bundle\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.195296 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hft4g\" (UniqueName: \"kubernetes.io/projected/87075836-5cb8-4014-b7f9-04c16c3e9e49-kube-api-access-hft4g\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.195314 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-fernet-keys\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.200551 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-scripts\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.201653 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-credential-keys\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.202224 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-config-data\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.206875 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-fernet-keys\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.209433 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-combined-ca-bundle\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.211173 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hft4g\" (UniqueName: \"kubernetes.io/projected/87075836-5cb8-4014-b7f9-04c16c3e9e49-kube-api-access-hft4g\") pod \"keystone-bootstrap-ztf2h\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.418203 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:09 crc kubenswrapper[4739]: E1201 15:53:09.458658 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 01 15:53:09 crc kubenswrapper[4739]: E1201 15:53:09.458922 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v8s7v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-g9q4b_openstack(588fd2a1-4b85-4af6-8b48-eff7fdf729ba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 15:53:09 crc kubenswrapper[4739]: E1201 15:53:09.460952 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-g9q4b" podUID="588fd2a1-4b85-4af6-8b48-eff7fdf729ba" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.546007 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-28jb8" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.548734 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.610744 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-config\") pod \"9ffc2832-c824-470f-8cfa-9ad4ea814125\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.610793 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-dns-svc\") pod \"9ffc2832-c824-470f-8cfa-9ad4ea814125\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.610832 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxr7t\" (UniqueName: \"kubernetes.io/projected/9ffc2832-c824-470f-8cfa-9ad4ea814125-kube-api-access-rxr7t\") pod \"9ffc2832-c824-470f-8cfa-9ad4ea814125\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.610855 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-ovsdbserver-nb\") pod \"9ffc2832-c824-470f-8cfa-9ad4ea814125\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.610902 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k844l\" (UniqueName: \"kubernetes.io/projected/c2999b95-023b-479c-a18c-0646e3e414d2-kube-api-access-k844l\") pod \"c2999b95-023b-479c-a18c-0646e3e414d2\" (UID: \"c2999b95-023b-479c-a18c-0646e3e414d2\") " Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.610951 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-ovsdbserver-sb\") pod \"9ffc2832-c824-470f-8cfa-9ad4ea814125\" (UID: \"9ffc2832-c824-470f-8cfa-9ad4ea814125\") " Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.610976 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2999b95-023b-479c-a18c-0646e3e414d2-combined-ca-bundle\") pod \"c2999b95-023b-479c-a18c-0646e3e414d2\" (UID: \"c2999b95-023b-479c-a18c-0646e3e414d2\") " Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.611026 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c2999b95-023b-479c-a18c-0646e3e414d2-config\") pod \"c2999b95-023b-479c-a18c-0646e3e414d2\" (UID: \"c2999b95-023b-479c-a18c-0646e3e414d2\") " Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.616436 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ffc2832-c824-470f-8cfa-9ad4ea814125-kube-api-access-rxr7t" (OuterVolumeSpecName: "kube-api-access-rxr7t") pod "9ffc2832-c824-470f-8cfa-9ad4ea814125" (UID: "9ffc2832-c824-470f-8cfa-9ad4ea814125"). InnerVolumeSpecName "kube-api-access-rxr7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.621577 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.621625 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.629804 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2999b95-023b-479c-a18c-0646e3e414d2-kube-api-access-k844l" (OuterVolumeSpecName: "kube-api-access-k844l") pod "c2999b95-023b-479c-a18c-0646e3e414d2" (UID: "c2999b95-023b-479c-a18c-0646e3e414d2"). InnerVolumeSpecName "kube-api-access-k844l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.646492 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2999b95-023b-479c-a18c-0646e3e414d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2999b95-023b-479c-a18c-0646e3e414d2" (UID: "c2999b95-023b-479c-a18c-0646e3e414d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.651205 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2999b95-023b-479c-a18c-0646e3e414d2-config" (OuterVolumeSpecName: "config") pod "c2999b95-023b-479c-a18c-0646e3e414d2" (UID: "c2999b95-023b-479c-a18c-0646e3e414d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.667764 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9ffc2832-c824-470f-8cfa-9ad4ea814125" (UID: "9ffc2832-c824-470f-8cfa-9ad4ea814125"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.672702 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9ffc2832-c824-470f-8cfa-9ad4ea814125" (UID: "9ffc2832-c824-470f-8cfa-9ad4ea814125"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.687196 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9ffc2832-c824-470f-8cfa-9ad4ea814125" (UID: "9ffc2832-c824-470f-8cfa-9ad4ea814125"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.705919 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-config" (OuterVolumeSpecName: "config") pod "9ffc2832-c824-470f-8cfa-9ad4ea814125" (UID: "9ffc2832-c824-470f-8cfa-9ad4ea814125"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.715448 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c2999b95-023b-479c-a18c-0646e3e414d2-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.715477 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.715486 4739 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.715496 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxr7t\" (UniqueName: \"kubernetes.io/projected/9ffc2832-c824-470f-8cfa-9ad4ea814125-kube-api-access-rxr7t\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.715508 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.715541 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k844l\" (UniqueName: \"kubernetes.io/projected/c2999b95-023b-479c-a18c-0646e3e414d2-kube-api-access-k844l\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.715549 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ffc2832-c824-470f-8cfa-9ad4ea814125-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.715558 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2999b95-023b-479c-a18c-0646e3e414d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.813878 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-28jb8" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.813892 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-28jb8" event={"ID":"c2999b95-023b-479c-a18c-0646e3e414d2","Type":"ContainerDied","Data":"5fb1f452d7536f7c64e149f9ab197834eef24d2d7530109abac42eda5ae5e594"} Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.813936 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fb1f452d7536f7c64e149f9ab197834eef24d2d7530109abac42eda5ae5e594" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.815893 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.815910 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" event={"ID":"9ffc2832-c824-470f-8cfa-9ad4ea814125","Type":"ContainerDied","Data":"a1db62b465838bf05e97955b72b236b49b13c059eeac9c511e2db4afd0f8c49e"} Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.815996 4739 scope.go:117] "RemoveContainer" containerID="9386ffe9422d13a240a8f5e843765882e844a274968e8d74e834de01f6d2566c" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.822979 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2lnl5" event={"ID":"baaebe35-daa8-449a-85fc-069056a1e9bc","Type":"ContainerStarted","Data":"97509525b652865b0169cfe3b320f3b871a74aae156451e1bbb456d5290ffaf4"} Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.829054 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f38003e3-7781-4602-89e0-0b89e7cf093b","Type":"ContainerStarted","Data":"6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a"} Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.850228 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-2lnl5" podStartSLOduration=1.918431511 podStartE2EDuration="22.850203586s" podCreationTimestamp="2025-12-01 15:52:47 +0000 UTC" firstStartedPulling="2025-12-01 15:52:48.430144916 +0000 UTC m=+1070.255891000" lastFinishedPulling="2025-12-01 15:53:09.361916981 +0000 UTC m=+1091.187663075" observedRunningTime="2025-12-01 15:53:09.850158905 +0000 UTC m=+1091.675905039" watchObservedRunningTime="2025-12-01 15:53:09.850203586 +0000 UTC m=+1091.675949680" Dec 01 15:53:09 crc kubenswrapper[4739]: E1201 15:53:09.858513 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-g9q4b" podUID="588fd2a1-4b85-4af6-8b48-eff7fdf729ba" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.858515 4739 scope.go:117] "RemoveContainer" containerID="974146f86ecb73d73747c3e21331e32fe0162b0168be30b462ed1f8dcc308199" Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.892445 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-nqkgc"] Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.902165 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54f9b7b8d9-nqkgc"] Dec 01 15:53:09 crc kubenswrapper[4739]: I1201 15:53:09.925857 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ztf2h"] Dec 01 15:53:09 crc kubenswrapper[4739]: W1201 15:53:09.929597 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87075836_5cb8_4014_b7f9_04c16c3e9e49.slice/crio-e9ba243295822bc3cadf7e037ec99ed9d8fd9a2ee0b7c28031a236bdb0e9afae WatchSource:0}: Error finding container e9ba243295822bc3cadf7e037ec99ed9d8fd9a2ee0b7c28031a236bdb0e9afae: Status 404 returned error can't find the container with id e9ba243295822bc3cadf7e037ec99ed9d8fd9a2ee0b7c28031a236bdb0e9afae Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.042541 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-vzh4g"] Dec 01 15:53:10 crc kubenswrapper[4739]: E1201 15:53:10.043351 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ffc2832-c824-470f-8cfa-9ad4ea814125" containerName="dnsmasq-dns" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.043378 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ffc2832-c824-470f-8cfa-9ad4ea814125" containerName="dnsmasq-dns" Dec 01 15:53:10 crc kubenswrapper[4739]: E1201 15:53:10.043394 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2999b95-023b-479c-a18c-0646e3e414d2" containerName="neutron-db-sync" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.043403 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2999b95-023b-479c-a18c-0646e3e414d2" containerName="neutron-db-sync" Dec 01 15:53:10 crc kubenswrapper[4739]: E1201 15:53:10.043448 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ffc2832-c824-470f-8cfa-9ad4ea814125" containerName="init" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.043459 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ffc2832-c824-470f-8cfa-9ad4ea814125" containerName="init" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.043684 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2999b95-023b-479c-a18c-0646e3e414d2" containerName="neutron-db-sync" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.043700 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ffc2832-c824-470f-8cfa-9ad4ea814125" containerName="dnsmasq-dns" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.044891 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.052786 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-vzh4g"] Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.110114 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7d77968bfd-m6mqb"] Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.111665 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.114259 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-r8k2p" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.119232 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.119613 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.120483 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.126008 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-ovsdbserver-sb\") pod \"dnsmasq-dns-7b946d459c-vzh4g\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.126220 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-httpd-config\") pod \"neutron-7d77968bfd-m6mqb\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.126339 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-config\") pod \"dnsmasq-dns-7b946d459c-vzh4g\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.126383 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d77968bfd-m6mqb"] Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.126551 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdv2c\" (UniqueName: \"kubernetes.io/projected/fa675bf0-e417-4847-ad0b-4f364208d007-kube-api-access-qdv2c\") pod \"neutron-7d77968bfd-m6mqb\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.126938 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-ovsdbserver-nb\") pod \"dnsmasq-dns-7b946d459c-vzh4g\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.127106 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8kxb\" (UniqueName: \"kubernetes.io/projected/9ce6139b-7b9f-449e-a642-ca71b4740873-kube-api-access-l8kxb\") pod \"dnsmasq-dns-7b946d459c-vzh4g\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.127259 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-config\") pod \"neutron-7d77968bfd-m6mqb\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.127450 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-combined-ca-bundle\") pod \"neutron-7d77968bfd-m6mqb\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.127721 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-ovndb-tls-certs\") pod \"neutron-7d77968bfd-m6mqb\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.127820 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-dns-svc\") pod \"dnsmasq-dns-7b946d459c-vzh4g\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.228379 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-ovsdbserver-nb\") pod \"dnsmasq-dns-7b946d459c-vzh4g\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.228441 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8kxb\" (UniqueName: \"kubernetes.io/projected/9ce6139b-7b9f-449e-a642-ca71b4740873-kube-api-access-l8kxb\") pod \"dnsmasq-dns-7b946d459c-vzh4g\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.228459 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-config\") pod \"neutron-7d77968bfd-m6mqb\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.228485 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-combined-ca-bundle\") pod \"neutron-7d77968bfd-m6mqb\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.228528 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-ovndb-tls-certs\") pod \"neutron-7d77968bfd-m6mqb\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.228544 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-dns-svc\") pod \"dnsmasq-dns-7b946d459c-vzh4g\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.228572 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-ovsdbserver-sb\") pod \"dnsmasq-dns-7b946d459c-vzh4g\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.228600 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-httpd-config\") pod \"neutron-7d77968bfd-m6mqb\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.228616 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-config\") pod \"dnsmasq-dns-7b946d459c-vzh4g\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.228631 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdv2c\" (UniqueName: \"kubernetes.io/projected/fa675bf0-e417-4847-ad0b-4f364208d007-kube-api-access-qdv2c\") pod \"neutron-7d77968bfd-m6mqb\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.229725 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-ovsdbserver-nb\") pod \"dnsmasq-dns-7b946d459c-vzh4g\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.231232 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-dns-svc\") pod \"dnsmasq-dns-7b946d459c-vzh4g\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.234929 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-config\") pod \"dnsmasq-dns-7b946d459c-vzh4g\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.235111 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-ovsdbserver-sb\") pod \"dnsmasq-dns-7b946d459c-vzh4g\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.237003 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-config\") pod \"neutron-7d77968bfd-m6mqb\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.239040 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-ovndb-tls-certs\") pod \"neutron-7d77968bfd-m6mqb\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.243330 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-combined-ca-bundle\") pod \"neutron-7d77968bfd-m6mqb\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.248780 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdv2c\" (UniqueName: \"kubernetes.io/projected/fa675bf0-e417-4847-ad0b-4f364208d007-kube-api-access-qdv2c\") pod \"neutron-7d77968bfd-m6mqb\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.249524 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-httpd-config\") pod \"neutron-7d77968bfd-m6mqb\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.279342 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8kxb\" (UniqueName: \"kubernetes.io/projected/9ce6139b-7b9f-449e-a642-ca71b4740873-kube-api-access-l8kxb\") pod \"dnsmasq-dns-7b946d459c-vzh4g\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.383171 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.433230 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.491464 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ffc2832-c824-470f-8cfa-9ad4ea814125" path="/var/lib/kubelet/pods/9ffc2832-c824-470f-8cfa-9ad4ea814125/volumes" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.492048 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1fca44a-582d-42af-978b-0b38945b260e" path="/var/lib/kubelet/pods/c1fca44a-582d-42af-978b-0b38945b260e/volumes" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.848549 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ztf2h" event={"ID":"87075836-5cb8-4014-b7f9-04c16c3e9e49","Type":"ContainerStarted","Data":"1d957ff4235e37c97e3da8c855c3964759b82846beb116b74642e3259af6fb86"} Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.848899 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ztf2h" event={"ID":"87075836-5cb8-4014-b7f9-04c16c3e9e49","Type":"ContainerStarted","Data":"e9ba243295822bc3cadf7e037ec99ed9d8fd9a2ee0b7c28031a236bdb0e9afae"} Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.865746 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-ztf2h" podStartSLOduration=1.8657248979999999 podStartE2EDuration="1.865724898s" podCreationTimestamp="2025-12-01 15:53:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:53:10.864780529 +0000 UTC m=+1092.690526633" watchObservedRunningTime="2025-12-01 15:53:10.865724898 +0000 UTC m=+1092.691470992" Dec 01 15:53:10 crc kubenswrapper[4739]: I1201 15:53:10.883669 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-vzh4g"] Dec 01 15:53:10 crc kubenswrapper[4739]: W1201 15:53:10.887446 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ce6139b_7b9f_449e_a642_ca71b4740873.slice/crio-61c498a63f17b128ac79c5ee5b1ae2dcb57527222698cc836f1e69beb3e75ad0 WatchSource:0}: Error finding container 61c498a63f17b128ac79c5ee5b1ae2dcb57527222698cc836f1e69beb3e75ad0: Status 404 returned error can't find the container with id 61c498a63f17b128ac79c5ee5b1ae2dcb57527222698cc836f1e69beb3e75ad0 Dec 01 15:53:11 crc kubenswrapper[4739]: I1201 15:53:11.147382 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d77968bfd-m6mqb"] Dec 01 15:53:11 crc kubenswrapper[4739]: W1201 15:53:11.337807 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa675bf0_e417_4847_ad0b_4f364208d007.slice/crio-8e4ab5bcb7ea05843314b1d96d8f250bd50647a4576ceedccfd7e6e146afb2b0 WatchSource:0}: Error finding container 8e4ab5bcb7ea05843314b1d96d8f250bd50647a4576ceedccfd7e6e146afb2b0: Status 404 returned error can't find the container with id 8e4ab5bcb7ea05843314b1d96d8f250bd50647a4576ceedccfd7e6e146afb2b0 Dec 01 15:53:11 crc kubenswrapper[4739]: I1201 15:53:11.866411 4739 generic.go:334] "Generic (PLEG): container finished" podID="baaebe35-daa8-449a-85fc-069056a1e9bc" containerID="97509525b652865b0169cfe3b320f3b871a74aae156451e1bbb456d5290ffaf4" exitCode=0 Dec 01 15:53:11 crc kubenswrapper[4739]: I1201 15:53:11.867011 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2lnl5" event={"ID":"baaebe35-daa8-449a-85fc-069056a1e9bc","Type":"ContainerDied","Data":"97509525b652865b0169cfe3b320f3b871a74aae156451e1bbb456d5290ffaf4"} Dec 01 15:53:11 crc kubenswrapper[4739]: I1201 15:53:11.878200 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d77968bfd-m6mqb" event={"ID":"fa675bf0-e417-4847-ad0b-4f364208d007","Type":"ContainerStarted","Data":"bfdf2eb822c5bb2516c0f817fac22d3ae840d3cdcfd06762d0f4823b9250636a"} Dec 01 15:53:11 crc kubenswrapper[4739]: I1201 15:53:11.878244 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d77968bfd-m6mqb" event={"ID":"fa675bf0-e417-4847-ad0b-4f364208d007","Type":"ContainerStarted","Data":"ba9b34bec7c662fef25c301e6be902f932f1082246a59b4dd662c2fc778b9706"} Dec 01 15:53:11 crc kubenswrapper[4739]: I1201 15:53:11.878253 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d77968bfd-m6mqb" event={"ID":"fa675bf0-e417-4847-ad0b-4f364208d007","Type":"ContainerStarted","Data":"8e4ab5bcb7ea05843314b1d96d8f250bd50647a4576ceedccfd7e6e146afb2b0"} Dec 01 15:53:11 crc kubenswrapper[4739]: I1201 15:53:11.879051 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:11 crc kubenswrapper[4739]: I1201 15:53:11.881677 4739 generic.go:334] "Generic (PLEG): container finished" podID="9ce6139b-7b9f-449e-a642-ca71b4740873" containerID="1d457b66af62af0d8203ec52864649b703d76cdf33f558e4d537cb02768b4bf0" exitCode=0 Dec 01 15:53:11 crc kubenswrapper[4739]: I1201 15:53:11.881745 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" event={"ID":"9ce6139b-7b9f-449e-a642-ca71b4740873","Type":"ContainerDied","Data":"1d457b66af62af0d8203ec52864649b703d76cdf33f558e4d537cb02768b4bf0"} Dec 01 15:53:11 crc kubenswrapper[4739]: I1201 15:53:11.881771 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" event={"ID":"9ce6139b-7b9f-449e-a642-ca71b4740873","Type":"ContainerStarted","Data":"61c498a63f17b128ac79c5ee5b1ae2dcb57527222698cc836f1e69beb3e75ad0"} Dec 01 15:53:11 crc kubenswrapper[4739]: I1201 15:53:11.891203 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f38003e3-7781-4602-89e0-0b89e7cf093b","Type":"ContainerStarted","Data":"d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56"} Dec 01 15:53:11 crc kubenswrapper[4739]: I1201 15:53:11.948586 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7d77968bfd-m6mqb" podStartSLOduration=1.948571209 podStartE2EDuration="1.948571209s" podCreationTimestamp="2025-12-01 15:53:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:53:11.945976919 +0000 UTC m=+1093.771723013" watchObservedRunningTime="2025-12-01 15:53:11.948571209 +0000 UTC m=+1093.774317303" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.606667 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d98bc475f-59p2r"] Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.609235 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.611916 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.618105 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.625020 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d98bc475f-59p2r"] Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.678834 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-ovndb-tls-certs\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.678886 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-internal-tls-certs\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.678933 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-httpd-config\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.678962 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-public-tls-certs\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.678979 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-combined-ca-bundle\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.679023 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqjz7\" (UniqueName: \"kubernetes.io/projected/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-kube-api-access-gqjz7\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.679052 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-config\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.781453 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqjz7\" (UniqueName: \"kubernetes.io/projected/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-kube-api-access-gqjz7\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.781523 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-config\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.781588 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-ovndb-tls-certs\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.781617 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-internal-tls-certs\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.781670 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-httpd-config\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.781706 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-public-tls-certs\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.781728 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-combined-ca-bundle\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.787110 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-internal-tls-certs\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.787110 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-public-tls-certs\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.787679 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-combined-ca-bundle\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.788169 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-ovndb-tls-certs\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.789377 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-config\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.797766 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-httpd-config\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.808512 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqjz7\" (UniqueName: \"kubernetes.io/projected/cb777746-9ae6-4eb5-8459-aa82ad2eafb5-kube-api-access-gqjz7\") pod \"neutron-5d98bc475f-59p2r\" (UID: \"cb777746-9ae6-4eb5-8459-aa82ad2eafb5\") " pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.902527 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" event={"ID":"9ce6139b-7b9f-449e-a642-ca71b4740873","Type":"ContainerStarted","Data":"e4e0e96b248781e6595254f727c731b71c43924c072648704ec9551a9e77679f"} Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.903080 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-54f9b7b8d9-nqkgc" podUID="9ffc2832-c824-470f-8cfa-9ad4ea814125" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: i/o timeout" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.929874 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" podStartSLOduration=2.929858154 podStartE2EDuration="2.929858154s" podCreationTimestamp="2025-12-01 15:53:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:53:12.927319686 +0000 UTC m=+1094.753065780" watchObservedRunningTime="2025-12-01 15:53:12.929858154 +0000 UTC m=+1094.755604248" Dec 01 15:53:12 crc kubenswrapper[4739]: I1201 15:53:12.932515 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.310561 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2lnl5" Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.391758 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmdhv\" (UniqueName: \"kubernetes.io/projected/baaebe35-daa8-449a-85fc-069056a1e9bc-kube-api-access-xmdhv\") pod \"baaebe35-daa8-449a-85fc-069056a1e9bc\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.391861 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/baaebe35-daa8-449a-85fc-069056a1e9bc-logs\") pod \"baaebe35-daa8-449a-85fc-069056a1e9bc\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.391928 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-combined-ca-bundle\") pod \"baaebe35-daa8-449a-85fc-069056a1e9bc\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.391945 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-config-data\") pod \"baaebe35-daa8-449a-85fc-069056a1e9bc\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.391996 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-scripts\") pod \"baaebe35-daa8-449a-85fc-069056a1e9bc\" (UID: \"baaebe35-daa8-449a-85fc-069056a1e9bc\") " Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.392209 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baaebe35-daa8-449a-85fc-069056a1e9bc-logs" (OuterVolumeSpecName: "logs") pod "baaebe35-daa8-449a-85fc-069056a1e9bc" (UID: "baaebe35-daa8-449a-85fc-069056a1e9bc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.392397 4739 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/baaebe35-daa8-449a-85fc-069056a1e9bc-logs\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.402175 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baaebe35-daa8-449a-85fc-069056a1e9bc-kube-api-access-xmdhv" (OuterVolumeSpecName: "kube-api-access-xmdhv") pod "baaebe35-daa8-449a-85fc-069056a1e9bc" (UID: "baaebe35-daa8-449a-85fc-069056a1e9bc"). InnerVolumeSpecName "kube-api-access-xmdhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.402180 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-scripts" (OuterVolumeSpecName: "scripts") pod "baaebe35-daa8-449a-85fc-069056a1e9bc" (UID: "baaebe35-daa8-449a-85fc-069056a1e9bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.415910 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-config-data" (OuterVolumeSpecName: "config-data") pod "baaebe35-daa8-449a-85fc-069056a1e9bc" (UID: "baaebe35-daa8-449a-85fc-069056a1e9bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.417586 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "baaebe35-daa8-449a-85fc-069056a1e9bc" (UID: "baaebe35-daa8-449a-85fc-069056a1e9bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.494015 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmdhv\" (UniqueName: \"kubernetes.io/projected/baaebe35-daa8-449a-85fc-069056a1e9bc-kube-api-access-xmdhv\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.494051 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.494092 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.494107 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baaebe35-daa8-449a-85fc-069056a1e9bc-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.527995 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d98bc475f-59p2r"] Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.913226 4739 generic.go:334] "Generic (PLEG): container finished" podID="87075836-5cb8-4014-b7f9-04c16c3e9e49" containerID="1d957ff4235e37c97e3da8c855c3964759b82846beb116b74642e3259af6fb86" exitCode=0 Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.913304 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ztf2h" event={"ID":"87075836-5cb8-4014-b7f9-04c16c3e9e49","Type":"ContainerDied","Data":"1d957ff4235e37c97e3da8c855c3964759b82846beb116b74642e3259af6fb86"} Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.919406 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d98bc475f-59p2r" event={"ID":"cb777746-9ae6-4eb5-8459-aa82ad2eafb5","Type":"ContainerStarted","Data":"19ebc24b8a6ca4e5eef8ba8ea5400c1d0e2863a69804f3ffa7c7cffa6454ad39"} Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.919468 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d98bc475f-59p2r" event={"ID":"cb777746-9ae6-4eb5-8459-aa82ad2eafb5","Type":"ContainerStarted","Data":"348d9b026176fdcb32e95bf9080c7cca7cc117b902e844582ccdd38bccef75ce"} Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.923074 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2lnl5" event={"ID":"baaebe35-daa8-449a-85fc-069056a1e9bc","Type":"ContainerDied","Data":"eb213b2c2ce901c1c6e5bbd2168ab8adda9b175bd0ef2b1f83857c5ca9113468"} Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.923123 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb213b2c2ce901c1c6e5bbd2168ab8adda9b175bd0ef2b1f83857c5ca9113468" Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.923225 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2lnl5" Dec 01 15:53:13 crc kubenswrapper[4739]: I1201 15:53:13.923344 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.018236 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-54db66b7f6-6plgv"] Dec 01 15:53:14 crc kubenswrapper[4739]: E1201 15:53:14.018974 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baaebe35-daa8-449a-85fc-069056a1e9bc" containerName="placement-db-sync" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.019117 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="baaebe35-daa8-449a-85fc-069056a1e9bc" containerName="placement-db-sync" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.019326 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="baaebe35-daa8-449a-85fc-069056a1e9bc" containerName="placement-db-sync" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.026730 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.028743 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-54db66b7f6-6plgv"] Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.044251 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.046835 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.050606 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-gxs8g" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.050831 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.050975 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.108322 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc9d7a70-2c80-4443-bd55-be94d972c336-public-tls-certs\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.108409 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc9d7a70-2c80-4443-bd55-be94d972c336-combined-ca-bundle\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.108548 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc9d7a70-2c80-4443-bd55-be94d972c336-internal-tls-certs\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.108595 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc9d7a70-2c80-4443-bd55-be94d972c336-logs\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.108662 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc9d7a70-2c80-4443-bd55-be94d972c336-config-data\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.108688 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9zmd\" (UniqueName: \"kubernetes.io/projected/cc9d7a70-2c80-4443-bd55-be94d972c336-kube-api-access-x9zmd\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.108705 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc9d7a70-2c80-4443-bd55-be94d972c336-scripts\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.211865 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc9d7a70-2c80-4443-bd55-be94d972c336-logs\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.211967 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc9d7a70-2c80-4443-bd55-be94d972c336-config-data\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.211998 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9zmd\" (UniqueName: \"kubernetes.io/projected/cc9d7a70-2c80-4443-bd55-be94d972c336-kube-api-access-x9zmd\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.212019 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc9d7a70-2c80-4443-bd55-be94d972c336-scripts\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.212137 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc9d7a70-2c80-4443-bd55-be94d972c336-public-tls-certs\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.212224 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc9d7a70-2c80-4443-bd55-be94d972c336-combined-ca-bundle\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.212267 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc9d7a70-2c80-4443-bd55-be94d972c336-internal-tls-certs\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.212438 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc9d7a70-2c80-4443-bd55-be94d972c336-logs\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.215972 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc9d7a70-2c80-4443-bd55-be94d972c336-config-data\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.216367 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc9d7a70-2c80-4443-bd55-be94d972c336-scripts\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.216634 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc9d7a70-2c80-4443-bd55-be94d972c336-combined-ca-bundle\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.217545 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc9d7a70-2c80-4443-bd55-be94d972c336-public-tls-certs\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.217854 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc9d7a70-2c80-4443-bd55-be94d972c336-internal-tls-certs\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.227379 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9zmd\" (UniqueName: \"kubernetes.io/projected/cc9d7a70-2c80-4443-bd55-be94d972c336-kube-api-access-x9zmd\") pod \"placement-54db66b7f6-6plgv\" (UID: \"cc9d7a70-2c80-4443-bd55-be94d972c336\") " pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.356753 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.934586 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d98bc475f-59p2r" event={"ID":"cb777746-9ae6-4eb5-8459-aa82ad2eafb5","Type":"ContainerStarted","Data":"59af56707a590d6bda36e1541112cbe1a73f2cc83c5216733e6f88fa0630d68a"} Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.935500 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:14 crc kubenswrapper[4739]: I1201 15:53:14.960209 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d98bc475f-59p2r" podStartSLOduration=2.960191957 podStartE2EDuration="2.960191957s" podCreationTimestamp="2025-12-01 15:53:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:53:14.956025958 +0000 UTC m=+1096.781772052" watchObservedRunningTime="2025-12-01 15:53:14.960191957 +0000 UTC m=+1096.785938051" Dec 01 15:53:19 crc kubenswrapper[4739]: I1201 15:53:19.867004 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:19.914216 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-credential-keys\") pod \"87075836-5cb8-4014-b7f9-04c16c3e9e49\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:19.914695 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-combined-ca-bundle\") pod \"87075836-5cb8-4014-b7f9-04c16c3e9e49\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:19.914742 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-config-data\") pod \"87075836-5cb8-4014-b7f9-04c16c3e9e49\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:19.914792 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-fernet-keys\") pod \"87075836-5cb8-4014-b7f9-04c16c3e9e49\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:19.914842 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hft4g\" (UniqueName: \"kubernetes.io/projected/87075836-5cb8-4014-b7f9-04c16c3e9e49-kube-api-access-hft4g\") pod \"87075836-5cb8-4014-b7f9-04c16c3e9e49\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:19.914906 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-scripts\") pod \"87075836-5cb8-4014-b7f9-04c16c3e9e49\" (UID: \"87075836-5cb8-4014-b7f9-04c16c3e9e49\") " Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:19.923836 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "87075836-5cb8-4014-b7f9-04c16c3e9e49" (UID: "87075836-5cb8-4014-b7f9-04c16c3e9e49"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:19.926577 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "87075836-5cb8-4014-b7f9-04c16c3e9e49" (UID: "87075836-5cb8-4014-b7f9-04c16c3e9e49"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:19.928406 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87075836-5cb8-4014-b7f9-04c16c3e9e49-kube-api-access-hft4g" (OuterVolumeSpecName: "kube-api-access-hft4g") pod "87075836-5cb8-4014-b7f9-04c16c3e9e49" (UID: "87075836-5cb8-4014-b7f9-04c16c3e9e49"). InnerVolumeSpecName "kube-api-access-hft4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:19.929678 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-scripts" (OuterVolumeSpecName: "scripts") pod "87075836-5cb8-4014-b7f9-04c16c3e9e49" (UID: "87075836-5cb8-4014-b7f9-04c16c3e9e49"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:19.952533 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87075836-5cb8-4014-b7f9-04c16c3e9e49" (UID: "87075836-5cb8-4014-b7f9-04c16c3e9e49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:19.977937 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-config-data" (OuterVolumeSpecName: "config-data") pod "87075836-5cb8-4014-b7f9-04c16c3e9e49" (UID: "87075836-5cb8-4014-b7f9-04c16c3e9e49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:20.017570 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hft4g\" (UniqueName: \"kubernetes.io/projected/87075836-5cb8-4014-b7f9-04c16c3e9e49-kube-api-access-hft4g\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:20.017602 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:20.017613 4739 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:20.017623 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:20.017634 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:20.017643 4739 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/87075836-5cb8-4014-b7f9-04c16c3e9e49-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:20.038731 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f38003e3-7781-4602-89e0-0b89e7cf093b","Type":"ContainerStarted","Data":"642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a"} Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:20.040936 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ztf2h" event={"ID":"87075836-5cb8-4014-b7f9-04c16c3e9e49","Type":"ContainerDied","Data":"e9ba243295822bc3cadf7e037ec99ed9d8fd9a2ee0b7c28031a236bdb0e9afae"} Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:20.040979 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9ba243295822bc3cadf7e037ec99ed9d8fd9a2ee0b7c28031a236bdb0e9afae" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:20.041025 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ztf2h" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:20.206836 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-54db66b7f6-6plgv"] Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:20.384752 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:20.459109 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-gz97r"] Dec 01 15:53:20 crc kubenswrapper[4739]: I1201 15:53:20.459771 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" podUID="6768159c-c437-4457-8d53-2032c1a9662e" containerName="dnsmasq-dns" containerID="cri-o://7ff5bc3a6f896fa2fcbf20d72dae92a9285a7ac2c8e6dccfa428ffd20842dd80" gracePeriod=10 Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.006349 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5f489c7bf6-cm54s"] Dec 01 15:53:21 crc kubenswrapper[4739]: E1201 15:53:21.007154 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87075836-5cb8-4014-b7f9-04c16c3e9e49" containerName="keystone-bootstrap" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.007173 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="87075836-5cb8-4014-b7f9-04c16c3e9e49" containerName="keystone-bootstrap" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.007386 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="87075836-5cb8-4014-b7f9-04c16c3e9e49" containerName="keystone-bootstrap" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.008280 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.010889 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zcbnm" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.011030 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.011088 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.010900 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.011256 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.011357 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.022544 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5f489c7bf6-cm54s"] Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.053743 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.080237 4739 generic.go:334] "Generic (PLEG): container finished" podID="6768159c-c437-4457-8d53-2032c1a9662e" containerID="7ff5bc3a6f896fa2fcbf20d72dae92a9285a7ac2c8e6dccfa428ffd20842dd80" exitCode=0 Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.080577 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.081261 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" event={"ID":"6768159c-c437-4457-8d53-2032c1a9662e","Type":"ContainerDied","Data":"7ff5bc3a6f896fa2fcbf20d72dae92a9285a7ac2c8e6dccfa428ffd20842dd80"} Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.081377 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987f74bbc-gz97r" event={"ID":"6768159c-c437-4457-8d53-2032c1a9662e","Type":"ContainerDied","Data":"7c04d2d730534cfa355a211cc1cefd2d94d2e7230e93b3b3357853c64d9aa3a5"} Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.081411 4739 scope.go:117] "RemoveContainer" containerID="7ff5bc3a6f896fa2fcbf20d72dae92a9285a7ac2c8e6dccfa428ffd20842dd80" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.101789 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-54db66b7f6-6plgv" event={"ID":"cc9d7a70-2c80-4443-bd55-be94d972c336","Type":"ContainerStarted","Data":"3831807467fb90e1a791d5ebece4e2e03b55c3e618dc636a64de249cd2f77e2b"} Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.101839 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-54db66b7f6-6plgv" event={"ID":"cc9d7a70-2c80-4443-bd55-be94d972c336","Type":"ContainerStarted","Data":"2da974836373849e28449757a8ebe4698b646a1f7ca582ad5d3a8867b6097654"} Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.101854 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-54db66b7f6-6plgv" event={"ID":"cc9d7a70-2c80-4443-bd55-be94d972c336","Type":"ContainerStarted","Data":"426487d57ce394068269c8375f77848fc277cd3d78914ea10b85c1c62ce04d96"} Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.102580 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.102639 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.125133 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-54db66b7f6-6plgv" podStartSLOduration=8.125108927 podStartE2EDuration="8.125108927s" podCreationTimestamp="2025-12-01 15:53:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:53:21.124650992 +0000 UTC m=+1102.950397086" watchObservedRunningTime="2025-12-01 15:53:21.125108927 +0000 UTC m=+1102.950855021" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.138312 4739 scope.go:117] "RemoveContainer" containerID="756a36be2e7180142ff4d2b3fc729d651f365c06a27ad8815c7af4f15cd1b634" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.138753 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-ovsdbserver-sb\") pod \"6768159c-c437-4457-8d53-2032c1a9662e\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.138910 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-ovsdbserver-nb\") pod \"6768159c-c437-4457-8d53-2032c1a9662e\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.138970 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7b5q\" (UniqueName: \"kubernetes.io/projected/6768159c-c437-4457-8d53-2032c1a9662e-kube-api-access-m7b5q\") pod \"6768159c-c437-4457-8d53-2032c1a9662e\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.139004 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-dns-svc\") pod \"6768159c-c437-4457-8d53-2032c1a9662e\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.139063 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-config\") pod \"6768159c-c437-4457-8d53-2032c1a9662e\" (UID: \"6768159c-c437-4457-8d53-2032c1a9662e\") " Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.139344 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-internal-tls-certs\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.139380 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-config-data\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.139433 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-fernet-keys\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.139455 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-combined-ca-bundle\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.139479 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-scripts\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.139545 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnxmw\" (UniqueName: \"kubernetes.io/projected/a244faee-6798-4c21-8041-6f5811a88eac-kube-api-access-pnxmw\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.139590 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-credential-keys\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.139627 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-public-tls-certs\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.168852 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6768159c-c437-4457-8d53-2032c1a9662e-kube-api-access-m7b5q" (OuterVolumeSpecName: "kube-api-access-m7b5q") pod "6768159c-c437-4457-8d53-2032c1a9662e" (UID: "6768159c-c437-4457-8d53-2032c1a9662e"). InnerVolumeSpecName "kube-api-access-m7b5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.203880 4739 scope.go:117] "RemoveContainer" containerID="7ff5bc3a6f896fa2fcbf20d72dae92a9285a7ac2c8e6dccfa428ffd20842dd80" Dec 01 15:53:21 crc kubenswrapper[4739]: E1201 15:53:21.205405 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ff5bc3a6f896fa2fcbf20d72dae92a9285a7ac2c8e6dccfa428ffd20842dd80\": container with ID starting with 7ff5bc3a6f896fa2fcbf20d72dae92a9285a7ac2c8e6dccfa428ffd20842dd80 not found: ID does not exist" containerID="7ff5bc3a6f896fa2fcbf20d72dae92a9285a7ac2c8e6dccfa428ffd20842dd80" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.205475 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ff5bc3a6f896fa2fcbf20d72dae92a9285a7ac2c8e6dccfa428ffd20842dd80"} err="failed to get container status \"7ff5bc3a6f896fa2fcbf20d72dae92a9285a7ac2c8e6dccfa428ffd20842dd80\": rpc error: code = NotFound desc = could not find container \"7ff5bc3a6f896fa2fcbf20d72dae92a9285a7ac2c8e6dccfa428ffd20842dd80\": container with ID starting with 7ff5bc3a6f896fa2fcbf20d72dae92a9285a7ac2c8e6dccfa428ffd20842dd80 not found: ID does not exist" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.205504 4739 scope.go:117] "RemoveContainer" containerID="756a36be2e7180142ff4d2b3fc729d651f365c06a27ad8815c7af4f15cd1b634" Dec 01 15:53:21 crc kubenswrapper[4739]: E1201 15:53:21.205879 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"756a36be2e7180142ff4d2b3fc729d651f365c06a27ad8815c7af4f15cd1b634\": container with ID starting with 756a36be2e7180142ff4d2b3fc729d651f365c06a27ad8815c7af4f15cd1b634 not found: ID does not exist" containerID="756a36be2e7180142ff4d2b3fc729d651f365c06a27ad8815c7af4f15cd1b634" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.205919 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"756a36be2e7180142ff4d2b3fc729d651f365c06a27ad8815c7af4f15cd1b634"} err="failed to get container status \"756a36be2e7180142ff4d2b3fc729d651f365c06a27ad8815c7af4f15cd1b634\": rpc error: code = NotFound desc = could not find container \"756a36be2e7180142ff4d2b3fc729d651f365c06a27ad8815c7af4f15cd1b634\": container with ID starting with 756a36be2e7180142ff4d2b3fc729d651f365c06a27ad8815c7af4f15cd1b634 not found: ID does not exist" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.212124 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6768159c-c437-4457-8d53-2032c1a9662e" (UID: "6768159c-c437-4457-8d53-2032c1a9662e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.225303 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6768159c-c437-4457-8d53-2032c1a9662e" (UID: "6768159c-c437-4457-8d53-2032c1a9662e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.226963 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-config" (OuterVolumeSpecName: "config") pod "6768159c-c437-4457-8d53-2032c1a9662e" (UID: "6768159c-c437-4457-8d53-2032c1a9662e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.228529 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6768159c-c437-4457-8d53-2032c1a9662e" (UID: "6768159c-c437-4457-8d53-2032c1a9662e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.240785 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-fernet-keys\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.240826 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-combined-ca-bundle\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.240861 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-scripts\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.240945 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnxmw\" (UniqueName: \"kubernetes.io/projected/a244faee-6798-4c21-8041-6f5811a88eac-kube-api-access-pnxmw\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.240982 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-credential-keys\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.241011 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-public-tls-certs\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.241034 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-internal-tls-certs\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.241051 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-config-data\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.241105 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.241115 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.241125 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7b5q\" (UniqueName: \"kubernetes.io/projected/6768159c-c437-4457-8d53-2032c1a9662e-kube-api-access-m7b5q\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.241135 4739 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.241144 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6768159c-c437-4457-8d53-2032c1a9662e-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.246564 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-fernet-keys\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.246585 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-combined-ca-bundle\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.247250 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-config-data\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.247715 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-internal-tls-certs\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.247981 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-scripts\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.248309 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-credential-keys\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.248666 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a244faee-6798-4c21-8041-6f5811a88eac-public-tls-certs\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.258662 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnxmw\" (UniqueName: \"kubernetes.io/projected/a244faee-6798-4c21-8041-6f5811a88eac-kube-api-access-pnxmw\") pod \"keystone-5f489c7bf6-cm54s\" (UID: \"a244faee-6798-4c21-8041-6f5811a88eac\") " pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.376136 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.427284 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-gz97r"] Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.436759 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7987f74bbc-gz97r"] Dec 01 15:53:21 crc kubenswrapper[4739]: I1201 15:53:21.862492 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5f489c7bf6-cm54s"] Dec 01 15:53:22 crc kubenswrapper[4739]: I1201 15:53:22.115312 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5f489c7bf6-cm54s" event={"ID":"a244faee-6798-4c21-8041-6f5811a88eac","Type":"ContainerStarted","Data":"b73c0661427cf1034580a18070c348f1736d787e800bd48a6e763f00e3ff9ced"} Dec 01 15:53:22 crc kubenswrapper[4739]: I1201 15:53:22.491250 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6768159c-c437-4457-8d53-2032c1a9662e" path="/var/lib/kubelet/pods/6768159c-c437-4457-8d53-2032c1a9662e/volumes" Dec 01 15:53:23 crc kubenswrapper[4739]: I1201 15:53:23.125583 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5f489c7bf6-cm54s" event={"ID":"a244faee-6798-4c21-8041-6f5811a88eac","Type":"ContainerStarted","Data":"d55e5ae648da8a9e438caaffdd98efe4643c1dd9d4159e6742eb6a3858da32bb"} Dec 01 15:53:23 crc kubenswrapper[4739]: I1201 15:53:23.125899 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:23 crc kubenswrapper[4739]: I1201 15:53:23.156153 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5f489c7bf6-cm54s" podStartSLOduration=3.156127359 podStartE2EDuration="3.156127359s" podCreationTimestamp="2025-12-01 15:53:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:53:23.149362881 +0000 UTC m=+1104.975108975" watchObservedRunningTime="2025-12-01 15:53:23.156127359 +0000 UTC m=+1104.981873453" Dec 01 15:53:24 crc kubenswrapper[4739]: I1201 15:53:24.141822 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-g9q4b" event={"ID":"588fd2a1-4b85-4af6-8b48-eff7fdf729ba","Type":"ContainerStarted","Data":"539633d585cc574c461499a29503013b1bac17403354c7b02502c66ae57990b4"} Dec 01 15:53:24 crc kubenswrapper[4739]: I1201 15:53:24.152053 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-kfsn8" event={"ID":"0cd4721c-4dfb-4219-9c69-c46299409e2a","Type":"ContainerStarted","Data":"09b40f88daf06ba6433f93b1e0b683e201d24da698cdacf14a17f83055f351a6"} Dec 01 15:53:24 crc kubenswrapper[4739]: I1201 15:53:24.172754 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-g9q4b" podStartSLOduration=2.190400542 podStartE2EDuration="37.172732895s" podCreationTimestamp="2025-12-01 15:52:47 +0000 UTC" firstStartedPulling="2025-12-01 15:52:48.305044861 +0000 UTC m=+1070.130790955" lastFinishedPulling="2025-12-01 15:53:23.287377204 +0000 UTC m=+1105.113123308" observedRunningTime="2025-12-01 15:53:24.167902456 +0000 UTC m=+1105.993648550" watchObservedRunningTime="2025-12-01 15:53:24.172732895 +0000 UTC m=+1105.998478989" Dec 01 15:53:24 crc kubenswrapper[4739]: I1201 15:53:24.195779 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-kfsn8" podStartSLOduration=2.340761228 podStartE2EDuration="37.195732776s" podCreationTimestamp="2025-12-01 15:52:47 +0000 UTC" firstStartedPulling="2025-12-01 15:52:48.393254296 +0000 UTC m=+1070.219000390" lastFinishedPulling="2025-12-01 15:53:23.248225854 +0000 UTC m=+1105.073971938" observedRunningTime="2025-12-01 15:53:24.190802944 +0000 UTC m=+1106.016549058" watchObservedRunningTime="2025-12-01 15:53:24.195732776 +0000 UTC m=+1106.021478870" Dec 01 15:53:27 crc kubenswrapper[4739]: I1201 15:53:27.187373 4739 generic.go:334] "Generic (PLEG): container finished" podID="0cd4721c-4dfb-4219-9c69-c46299409e2a" containerID="09b40f88daf06ba6433f93b1e0b683e201d24da698cdacf14a17f83055f351a6" exitCode=0 Dec 01 15:53:27 crc kubenswrapper[4739]: I1201 15:53:27.187475 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-kfsn8" event={"ID":"0cd4721c-4dfb-4219-9c69-c46299409e2a","Type":"ContainerDied","Data":"09b40f88daf06ba6433f93b1e0b683e201d24da698cdacf14a17f83055f351a6"} Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.243061 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f38003e3-7781-4602-89e0-0b89e7cf093b","Type":"ContainerStarted","Data":"98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338"} Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.243321 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="ceilometer-central-agent" containerID="cri-o://6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a" gracePeriod=30 Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.243457 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="proxy-httpd" containerID="cri-o://98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338" gracePeriod=30 Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.243511 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="sg-core" containerID="cri-o://642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a" gracePeriod=30 Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.243552 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="ceilometer-notification-agent" containerID="cri-o://d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56" gracePeriod=30 Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.289670 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.400536724 podStartE2EDuration="42.289648017s" podCreationTimestamp="2025-12-01 15:52:46 +0000 UTC" firstStartedPulling="2025-12-01 15:52:48.235176333 +0000 UTC m=+1070.060922427" lastFinishedPulling="2025-12-01 15:53:27.124287626 +0000 UTC m=+1108.950033720" observedRunningTime="2025-12-01 15:53:28.274956444 +0000 UTC m=+1110.100702558" watchObservedRunningTime="2025-12-01 15:53:28.289648017 +0000 UTC m=+1110.115394131" Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.657536 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-kfsn8" Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.811150 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cd4721c-4dfb-4219-9c69-c46299409e2a-combined-ca-bundle\") pod \"0cd4721c-4dfb-4219-9c69-c46299409e2a\" (UID: \"0cd4721c-4dfb-4219-9c69-c46299409e2a\") " Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.811292 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0cd4721c-4dfb-4219-9c69-c46299409e2a-db-sync-config-data\") pod \"0cd4721c-4dfb-4219-9c69-c46299409e2a\" (UID: \"0cd4721c-4dfb-4219-9c69-c46299409e2a\") " Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.811598 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jk9x\" (UniqueName: \"kubernetes.io/projected/0cd4721c-4dfb-4219-9c69-c46299409e2a-kube-api-access-4jk9x\") pod \"0cd4721c-4dfb-4219-9c69-c46299409e2a\" (UID: \"0cd4721c-4dfb-4219-9c69-c46299409e2a\") " Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.817318 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cd4721c-4dfb-4219-9c69-c46299409e2a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0cd4721c-4dfb-4219-9c69-c46299409e2a" (UID: "0cd4721c-4dfb-4219-9c69-c46299409e2a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.817632 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cd4721c-4dfb-4219-9c69-c46299409e2a-kube-api-access-4jk9x" (OuterVolumeSpecName: "kube-api-access-4jk9x") pod "0cd4721c-4dfb-4219-9c69-c46299409e2a" (UID: "0cd4721c-4dfb-4219-9c69-c46299409e2a"). InnerVolumeSpecName "kube-api-access-4jk9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.835506 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cd4721c-4dfb-4219-9c69-c46299409e2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0cd4721c-4dfb-4219-9c69-c46299409e2a" (UID: "0cd4721c-4dfb-4219-9c69-c46299409e2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.915218 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jk9x\" (UniqueName: \"kubernetes.io/projected/0cd4721c-4dfb-4219-9c69-c46299409e2a-kube-api-access-4jk9x\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.915249 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cd4721c-4dfb-4219-9c69-c46299409e2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:28 crc kubenswrapper[4739]: I1201 15:53:28.915260 4739 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0cd4721c-4dfb-4219-9c69-c46299409e2a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.114138 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.219238 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbp5v\" (UniqueName: \"kubernetes.io/projected/f38003e3-7781-4602-89e0-0b89e7cf093b-kube-api-access-vbp5v\") pod \"f38003e3-7781-4602-89e0-0b89e7cf093b\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.219314 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-scripts\") pod \"f38003e3-7781-4602-89e0-0b89e7cf093b\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.219348 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f38003e3-7781-4602-89e0-0b89e7cf093b-run-httpd\") pod \"f38003e3-7781-4602-89e0-0b89e7cf093b\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.219966 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f38003e3-7781-4602-89e0-0b89e7cf093b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f38003e3-7781-4602-89e0-0b89e7cf093b" (UID: "f38003e3-7781-4602-89e0-0b89e7cf093b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.219549 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f38003e3-7781-4602-89e0-0b89e7cf093b-log-httpd\") pod \"f38003e3-7781-4602-89e0-0b89e7cf093b\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.220161 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f38003e3-7781-4602-89e0-0b89e7cf093b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f38003e3-7781-4602-89e0-0b89e7cf093b" (UID: "f38003e3-7781-4602-89e0-0b89e7cf093b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.220188 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-config-data\") pod \"f38003e3-7781-4602-89e0-0b89e7cf093b\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.220476 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-combined-ca-bundle\") pod \"f38003e3-7781-4602-89e0-0b89e7cf093b\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.220553 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-sg-core-conf-yaml\") pod \"f38003e3-7781-4602-89e0-0b89e7cf093b\" (UID: \"f38003e3-7781-4602-89e0-0b89e7cf093b\") " Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.221224 4739 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f38003e3-7781-4602-89e0-0b89e7cf093b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.221247 4739 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f38003e3-7781-4602-89e0-0b89e7cf093b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.223544 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-scripts" (OuterVolumeSpecName: "scripts") pod "f38003e3-7781-4602-89e0-0b89e7cf093b" (UID: "f38003e3-7781-4602-89e0-0b89e7cf093b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.225675 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f38003e3-7781-4602-89e0-0b89e7cf093b-kube-api-access-vbp5v" (OuterVolumeSpecName: "kube-api-access-vbp5v") pod "f38003e3-7781-4602-89e0-0b89e7cf093b" (UID: "f38003e3-7781-4602-89e0-0b89e7cf093b"). InnerVolumeSpecName "kube-api-access-vbp5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.267822 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-kfsn8" event={"ID":"0cd4721c-4dfb-4219-9c69-c46299409e2a","Type":"ContainerDied","Data":"989c4525a84b9c9ad7400f582965fd696508e84251081679096e077f8819060d"} Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.267894 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="989c4525a84b9c9ad7400f582965fd696508e84251081679096e077f8819060d" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.267987 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-kfsn8" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.275235 4739 generic.go:334] "Generic (PLEG): container finished" podID="588fd2a1-4b85-4af6-8b48-eff7fdf729ba" containerID="539633d585cc574c461499a29503013b1bac17403354c7b02502c66ae57990b4" exitCode=0 Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.275367 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-g9q4b" event={"ID":"588fd2a1-4b85-4af6-8b48-eff7fdf729ba","Type":"ContainerDied","Data":"539633d585cc574c461499a29503013b1bac17403354c7b02502c66ae57990b4"} Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.276367 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f38003e3-7781-4602-89e0-0b89e7cf093b" (UID: "f38003e3-7781-4602-89e0-0b89e7cf093b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.280793 4739 generic.go:334] "Generic (PLEG): container finished" podID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerID="98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338" exitCode=0 Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.280828 4739 generic.go:334] "Generic (PLEG): container finished" podID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerID="642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a" exitCode=2 Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.280835 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.280848 4739 generic.go:334] "Generic (PLEG): container finished" podID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerID="d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56" exitCode=0 Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.280864 4739 generic.go:334] "Generic (PLEG): container finished" podID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerID="6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a" exitCode=0 Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.280893 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f38003e3-7781-4602-89e0-0b89e7cf093b","Type":"ContainerDied","Data":"98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338"} Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.280932 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f38003e3-7781-4602-89e0-0b89e7cf093b","Type":"ContainerDied","Data":"642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a"} Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.280952 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f38003e3-7781-4602-89e0-0b89e7cf093b","Type":"ContainerDied","Data":"d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56"} Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.280970 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f38003e3-7781-4602-89e0-0b89e7cf093b","Type":"ContainerDied","Data":"6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a"} Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.280988 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f38003e3-7781-4602-89e0-0b89e7cf093b","Type":"ContainerDied","Data":"7841d2ec29d1b4929d6f3fa922532bc9481553b1f8303470a319c4e4ee4f752f"} Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.281035 4739 scope.go:117] "RemoveContainer" containerID="98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.306610 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f38003e3-7781-4602-89e0-0b89e7cf093b" (UID: "f38003e3-7781-4602-89e0-0b89e7cf093b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.314031 4739 scope.go:117] "RemoveContainer" containerID="642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.322522 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbp5v\" (UniqueName: \"kubernetes.io/projected/f38003e3-7781-4602-89e0-0b89e7cf093b-kube-api-access-vbp5v\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.322557 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.322571 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.322582 4739 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.333500 4739 scope.go:117] "RemoveContainer" containerID="d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.351511 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-config-data" (OuterVolumeSpecName: "config-data") pod "f38003e3-7781-4602-89e0-0b89e7cf093b" (UID: "f38003e3-7781-4602-89e0-0b89e7cf093b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.362269 4739 scope.go:117] "RemoveContainer" containerID="6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.424675 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f38003e3-7781-4602-89e0-0b89e7cf093b-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.450450 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6bbd6b9f5c-wjbz6"] Dec 01 15:53:29 crc kubenswrapper[4739]: E1201 15:53:29.450922 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="sg-core" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.450941 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="sg-core" Dec 01 15:53:29 crc kubenswrapper[4739]: E1201 15:53:29.450961 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="ceilometer-central-agent" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.450975 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="ceilometer-central-agent" Dec 01 15:53:29 crc kubenswrapper[4739]: E1201 15:53:29.451002 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6768159c-c437-4457-8d53-2032c1a9662e" containerName="init" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.451014 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="6768159c-c437-4457-8d53-2032c1a9662e" containerName="init" Dec 01 15:53:29 crc kubenswrapper[4739]: E1201 15:53:29.451045 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6768159c-c437-4457-8d53-2032c1a9662e" containerName="dnsmasq-dns" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.451057 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="6768159c-c437-4457-8d53-2032c1a9662e" containerName="dnsmasq-dns" Dec 01 15:53:29 crc kubenswrapper[4739]: E1201 15:53:29.451083 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="proxy-httpd" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.451099 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="proxy-httpd" Dec 01 15:53:29 crc kubenswrapper[4739]: E1201 15:53:29.451122 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="ceilometer-notification-agent" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.451139 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="ceilometer-notification-agent" Dec 01 15:53:29 crc kubenswrapper[4739]: E1201 15:53:29.451228 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cd4721c-4dfb-4219-9c69-c46299409e2a" containerName="barbican-db-sync" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.451248 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cd4721c-4dfb-4219-9c69-c46299409e2a" containerName="barbican-db-sync" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.457598 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="ceilometer-notification-agent" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.457968 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="ceilometer-central-agent" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.457999 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cd4721c-4dfb-4219-9c69-c46299409e2a" containerName="barbican-db-sync" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.458027 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="sg-core" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.458055 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" containerName="proxy-httpd" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.458103 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="6768159c-c437-4457-8d53-2032c1a9662e" containerName="dnsmasq-dns" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.459707 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.462168 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-g66cf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.462202 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.463030 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.477178 4739 scope.go:117] "RemoveContainer" containerID="98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338" Dec 01 15:53:29 crc kubenswrapper[4739]: E1201 15:53:29.478550 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338\": container with ID starting with 98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338 not found: ID does not exist" containerID="98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.478611 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338"} err="failed to get container status \"98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338\": rpc error: code = NotFound desc = could not find container \"98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338\": container with ID starting with 98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338 not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.478649 4739 scope.go:117] "RemoveContainer" containerID="642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a" Dec 01 15:53:29 crc kubenswrapper[4739]: E1201 15:53:29.479923 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a\": container with ID starting with 642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a not found: ID does not exist" containerID="642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.479955 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a"} err="failed to get container status \"642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a\": rpc error: code = NotFound desc = could not find container \"642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a\": container with ID starting with 642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.479985 4739 scope.go:117] "RemoveContainer" containerID="d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56" Dec 01 15:53:29 crc kubenswrapper[4739]: E1201 15:53:29.482356 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56\": container with ID starting with d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56 not found: ID does not exist" containerID="d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.482443 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56"} err="failed to get container status \"d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56\": rpc error: code = NotFound desc = could not find container \"d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56\": container with ID starting with d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56 not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.482483 4739 scope.go:117] "RemoveContainer" containerID="6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a" Dec 01 15:53:29 crc kubenswrapper[4739]: E1201 15:53:29.492600 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a\": container with ID starting with 6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a not found: ID does not exist" containerID="6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.492660 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a"} err="failed to get container status \"6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a\": rpc error: code = NotFound desc = could not find container \"6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a\": container with ID starting with 6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.492696 4739 scope.go:117] "RemoveContainer" containerID="98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.492901 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6bbd6b9f5c-wjbz6"] Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.493985 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338"} err="failed to get container status \"98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338\": rpc error: code = NotFound desc = could not find container \"98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338\": container with ID starting with 98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338 not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.494032 4739 scope.go:117] "RemoveContainer" containerID="642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.494514 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a"} err="failed to get container status \"642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a\": rpc error: code = NotFound desc = could not find container \"642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a\": container with ID starting with 642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.494559 4739 scope.go:117] "RemoveContainer" containerID="d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.495729 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56"} err="failed to get container status \"d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56\": rpc error: code = NotFound desc = could not find container \"d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56\": container with ID starting with d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56 not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.495762 4739 scope.go:117] "RemoveContainer" containerID="6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.496037 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a"} err="failed to get container status \"6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a\": rpc error: code = NotFound desc = could not find container \"6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a\": container with ID starting with 6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.496060 4739 scope.go:117] "RemoveContainer" containerID="98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.496279 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338"} err="failed to get container status \"98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338\": rpc error: code = NotFound desc = could not find container \"98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338\": container with ID starting with 98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338 not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.496304 4739 scope.go:117] "RemoveContainer" containerID="642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.496528 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a"} err="failed to get container status \"642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a\": rpc error: code = NotFound desc = could not find container \"642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a\": container with ID starting with 642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.496553 4739 scope.go:117] "RemoveContainer" containerID="d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.496754 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56"} err="failed to get container status \"d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56\": rpc error: code = NotFound desc = could not find container \"d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56\": container with ID starting with d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56 not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.496780 4739 scope.go:117] "RemoveContainer" containerID="6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.497571 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a"} err="failed to get container status \"6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a\": rpc error: code = NotFound desc = could not find container \"6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a\": container with ID starting with 6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.497606 4739 scope.go:117] "RemoveContainer" containerID="98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.502930 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338"} err="failed to get container status \"98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338\": rpc error: code = NotFound desc = could not find container \"98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338\": container with ID starting with 98d986e95f4f1741c4aa8233e4099b47c7e85e6b230cbeba68ce0d6481c41338 not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.502972 4739 scope.go:117] "RemoveContainer" containerID="642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.503316 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a"} err="failed to get container status \"642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a\": rpc error: code = NotFound desc = could not find container \"642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a\": container with ID starting with 642a905016f4503abab282ce32a7fb4e3d419b20fd5dd477edf0cd2b26e8646a not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.503356 4739 scope.go:117] "RemoveContainer" containerID="d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.504870 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56"} err="failed to get container status \"d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56\": rpc error: code = NotFound desc = could not find container \"d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56\": container with ID starting with d7077cc239e575a249f74b075de6981007e4cf4535e8c97ff87d837835ba1b56 not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.504919 4739 scope.go:117] "RemoveContainer" containerID="6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.505531 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a"} err="failed to get container status \"6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a\": rpc error: code = NotFound desc = could not find container \"6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a\": container with ID starting with 6afe39e7395bdd08742ef2c09c20d1d31771892075ca248650198dbbd1f6e84a not found: ID does not exist" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.536692 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-ff455997d-9lg5c"] Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.537915 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.543438 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.559570 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-ff455997d-9lg5c"] Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.609856 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-cdphf"] Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.611244 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.626843 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f9d262-1e44-42e4-a413-b85bb5f25241-config-data\") pod \"barbican-worker-6bbd6b9f5c-wjbz6\" (UID: \"02f9d262-1e44-42e4-a413-b85bb5f25241\") " pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.626877 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02f9d262-1e44-42e4-a413-b85bb5f25241-config-data-custom\") pod \"barbican-worker-6bbd6b9f5c-wjbz6\" (UID: \"02f9d262-1e44-42e4-a413-b85bb5f25241\") " pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.626919 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd-config-data\") pod \"barbican-keystone-listener-ff455997d-9lg5c\" (UID: \"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd\") " pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.626944 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crh2m\" (UniqueName: \"kubernetes.io/projected/2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd-kube-api-access-crh2m\") pod \"barbican-keystone-listener-ff455997d-9lg5c\" (UID: \"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd\") " pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.626964 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4f4t\" (UniqueName: \"kubernetes.io/projected/02f9d262-1e44-42e4-a413-b85bb5f25241-kube-api-access-z4f4t\") pod \"barbican-worker-6bbd6b9f5c-wjbz6\" (UID: \"02f9d262-1e44-42e4-a413-b85bb5f25241\") " pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.627025 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd-combined-ca-bundle\") pod \"barbican-keystone-listener-ff455997d-9lg5c\" (UID: \"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd\") " pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.627042 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd-config-data-custom\") pod \"barbican-keystone-listener-ff455997d-9lg5c\" (UID: \"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd\") " pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.627078 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02f9d262-1e44-42e4-a413-b85bb5f25241-logs\") pod \"barbican-worker-6bbd6b9f5c-wjbz6\" (UID: \"02f9d262-1e44-42e4-a413-b85bb5f25241\") " pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.627129 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd-logs\") pod \"barbican-keystone-listener-ff455997d-9lg5c\" (UID: \"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd\") " pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.627146 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f9d262-1e44-42e4-a413-b85bb5f25241-combined-ca-bundle\") pod \"barbican-worker-6bbd6b9f5c-wjbz6\" (UID: \"02f9d262-1e44-42e4-a413-b85bb5f25241\") " pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.627270 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-cdphf"] Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.643476 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.665493 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.682231 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.684196 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.686080 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.691860 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.723471 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.730030 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f9d262-1e44-42e4-a413-b85bb5f25241-config-data\") pod \"barbican-worker-6bbd6b9f5c-wjbz6\" (UID: \"02f9d262-1e44-42e4-a413-b85bb5f25241\") " pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.730070 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02f9d262-1e44-42e4-a413-b85bb5f25241-config-data-custom\") pod \"barbican-worker-6bbd6b9f5c-wjbz6\" (UID: \"02f9d262-1e44-42e4-a413-b85bb5f25241\") " pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.730143 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd-config-data\") pod \"barbican-keystone-listener-ff455997d-9lg5c\" (UID: \"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd\") " pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.730161 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crh2m\" (UniqueName: \"kubernetes.io/projected/2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd-kube-api-access-crh2m\") pod \"barbican-keystone-listener-ff455997d-9lg5c\" (UID: \"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd\") " pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.730186 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4f4t\" (UniqueName: \"kubernetes.io/projected/02f9d262-1e44-42e4-a413-b85bb5f25241-kube-api-access-z4f4t\") pod \"barbican-worker-6bbd6b9f5c-wjbz6\" (UID: \"02f9d262-1e44-42e4-a413-b85bb5f25241\") " pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.730232 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb684768f-cdphf\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.730315 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb684768f-cdphf\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.730340 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd-config-data-custom\") pod \"barbican-keystone-listener-ff455997d-9lg5c\" (UID: \"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd\") " pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.730357 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd-combined-ca-bundle\") pod \"barbican-keystone-listener-ff455997d-9lg5c\" (UID: \"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd\") " pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.730409 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02f9d262-1e44-42e4-a413-b85bb5f25241-logs\") pod \"barbican-worker-6bbd6b9f5c-wjbz6\" (UID: \"02f9d262-1e44-42e4-a413-b85bb5f25241\") " pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.730441 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-dns-svc\") pod \"dnsmasq-dns-6bb684768f-cdphf\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.730471 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd-logs\") pod \"barbican-keystone-listener-ff455997d-9lg5c\" (UID: \"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd\") " pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.730488 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f9d262-1e44-42e4-a413-b85bb5f25241-combined-ca-bundle\") pod \"barbican-worker-6bbd6b9f5c-wjbz6\" (UID: \"02f9d262-1e44-42e4-a413-b85bb5f25241\") " pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.730510 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8gzt\" (UniqueName: \"kubernetes.io/projected/679c7628-a662-4085-8fd9-ae49692f674c-kube-api-access-k8gzt\") pod \"dnsmasq-dns-6bb684768f-cdphf\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.730556 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-config\") pod \"dnsmasq-dns-6bb684768f-cdphf\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.733971 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02f9d262-1e44-42e4-a413-b85bb5f25241-logs\") pod \"barbican-worker-6bbd6b9f5c-wjbz6\" (UID: \"02f9d262-1e44-42e4-a413-b85bb5f25241\") " pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.734280 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd-logs\") pod \"barbican-keystone-listener-ff455997d-9lg5c\" (UID: \"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd\") " pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.734907 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-799f5fcdc4-bwhj9"] Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.735409 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd-combined-ca-bundle\") pod \"barbican-keystone-listener-ff455997d-9lg5c\" (UID: \"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd\") " pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.735535 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02f9d262-1e44-42e4-a413-b85bb5f25241-config-data-custom\") pod \"barbican-worker-6bbd6b9f5c-wjbz6\" (UID: \"02f9d262-1e44-42e4-a413-b85bb5f25241\") " pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.736766 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f9d262-1e44-42e4-a413-b85bb5f25241-combined-ca-bundle\") pod \"barbican-worker-6bbd6b9f5c-wjbz6\" (UID: \"02f9d262-1e44-42e4-a413-b85bb5f25241\") " pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.739378 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f9d262-1e44-42e4-a413-b85bb5f25241-config-data\") pod \"barbican-worker-6bbd6b9f5c-wjbz6\" (UID: \"02f9d262-1e44-42e4-a413-b85bb5f25241\") " pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.739475 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.749235 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-799f5fcdc4-bwhj9"] Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.750157 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd-config-data\") pod \"barbican-keystone-listener-ff455997d-9lg5c\" (UID: \"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd\") " pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.754226 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd-config-data-custom\") pod \"barbican-keystone-listener-ff455997d-9lg5c\" (UID: \"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd\") " pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.754334 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.758109 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crh2m\" (UniqueName: \"kubernetes.io/projected/2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd-kube-api-access-crh2m\") pod \"barbican-keystone-listener-ff455997d-9lg5c\" (UID: \"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd\") " pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.760045 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4f4t\" (UniqueName: \"kubernetes.io/projected/02f9d262-1e44-42e4-a413-b85bb5f25241-kube-api-access-z4f4t\") pod \"barbican-worker-6bbd6b9f5c-wjbz6\" (UID: \"02f9d262-1e44-42e4-a413-b85bb5f25241\") " pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.775566 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.831663 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r669x\" (UniqueName: \"kubernetes.io/projected/d79c5700-e1a6-416b-bba1-5c211439c9d9-kube-api-access-r669x\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.831714 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.831757 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb684768f-cdphf\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.831785 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.831802 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d79c5700-e1a6-416b-bba1-5c211439c9d9-run-httpd\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.831824 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-scripts\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.831843 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d79c5700-e1a6-416b-bba1-5c211439c9d9-log-httpd\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.831859 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-config-data\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.831879 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb684768f-cdphf\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.831904 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3703361d-b3f1-4139-85da-cf21e825b7f2-logs\") pod \"barbican-api-799f5fcdc4-bwhj9\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.831934 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-dns-svc\") pod \"dnsmasq-dns-6bb684768f-cdphf\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.831953 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgpvg\" (UniqueName: \"kubernetes.io/projected/3703361d-b3f1-4139-85da-cf21e825b7f2-kube-api-access-pgpvg\") pod \"barbican-api-799f5fcdc4-bwhj9\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.831975 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8gzt\" (UniqueName: \"kubernetes.io/projected/679c7628-a662-4085-8fd9-ae49692f674c-kube-api-access-k8gzt\") pod \"dnsmasq-dns-6bb684768f-cdphf\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.831996 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-combined-ca-bundle\") pod \"barbican-api-799f5fcdc4-bwhj9\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.832020 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-config\") pod \"dnsmasq-dns-6bb684768f-cdphf\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.832036 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-config-data\") pod \"barbican-api-799f5fcdc4-bwhj9\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.832059 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-config-data-custom\") pod \"barbican-api-799f5fcdc4-bwhj9\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.832954 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-config\") pod \"dnsmasq-dns-6bb684768f-cdphf\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.833059 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb684768f-cdphf\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.833097 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb684768f-cdphf\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.834180 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-dns-svc\") pod \"dnsmasq-dns-6bb684768f-cdphf\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.851876 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8gzt\" (UniqueName: \"kubernetes.io/projected/679c7628-a662-4085-8fd9-ae49692f674c-kube-api-access-k8gzt\") pod \"dnsmasq-dns-6bb684768f-cdphf\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.855055 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.933173 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.933815 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.933847 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d79c5700-e1a6-416b-bba1-5c211439c9d9-run-httpd\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.933882 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-scripts\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.933909 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d79c5700-e1a6-416b-bba1-5c211439c9d9-log-httpd\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.933931 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-config-data\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.933966 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3703361d-b3f1-4139-85da-cf21e825b7f2-logs\") pod \"barbican-api-799f5fcdc4-bwhj9\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.934014 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgpvg\" (UniqueName: \"kubernetes.io/projected/3703361d-b3f1-4139-85da-cf21e825b7f2-kube-api-access-pgpvg\") pod \"barbican-api-799f5fcdc4-bwhj9\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.934047 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-combined-ca-bundle\") pod \"barbican-api-799f5fcdc4-bwhj9\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.934091 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-config-data\") pod \"barbican-api-799f5fcdc4-bwhj9\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.934124 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-config-data-custom\") pod \"barbican-api-799f5fcdc4-bwhj9\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.934170 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r669x\" (UniqueName: \"kubernetes.io/projected/d79c5700-e1a6-416b-bba1-5c211439c9d9-kube-api-access-r669x\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.934196 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.934397 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d79c5700-e1a6-416b-bba1-5c211439c9d9-run-httpd\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.934449 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d79c5700-e1a6-416b-bba1-5c211439c9d9-log-httpd\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.935270 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3703361d-b3f1-4139-85da-cf21e825b7f2-logs\") pod \"barbican-api-799f5fcdc4-bwhj9\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.939786 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.939854 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-combined-ca-bundle\") pod \"barbican-api-799f5fcdc4-bwhj9\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.940076 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.941761 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-config-data\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.942897 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-config-data-custom\") pod \"barbican-api-799f5fcdc4-bwhj9\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.943189 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-config-data\") pod \"barbican-api-799f5fcdc4-bwhj9\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.946402 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-scripts\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.951545 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgpvg\" (UniqueName: \"kubernetes.io/projected/3703361d-b3f1-4139-85da-cf21e825b7f2-kube-api-access-pgpvg\") pod \"barbican-api-799f5fcdc4-bwhj9\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:29 crc kubenswrapper[4739]: I1201 15:53:29.954989 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r669x\" (UniqueName: \"kubernetes.io/projected/d79c5700-e1a6-416b-bba1-5c211439c9d9-kube-api-access-r669x\") pod \"ceilometer-0\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " pod="openstack/ceilometer-0" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.015850 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.162859 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.206286 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6bbd6b9f5c-wjbz6"] Dec 01 15:53:30 crc kubenswrapper[4739]: W1201 15:53:30.217182 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02f9d262_1e44_42e4_a413_b85bb5f25241.slice/crio-da55a126f8050c522c35413222d4078d6d32abcfafb27e1231896811da61d931 WatchSource:0}: Error finding container da55a126f8050c522c35413222d4078d6d32abcfafb27e1231896811da61d931: Status 404 returned error can't find the container with id da55a126f8050c522c35413222d4078d6d32abcfafb27e1231896811da61d931 Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.294076 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" event={"ID":"02f9d262-1e44-42e4-a413-b85bb5f25241","Type":"ContainerStarted","Data":"da55a126f8050c522c35413222d4078d6d32abcfafb27e1231896811da61d931"} Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.300845 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-ff455997d-9lg5c"] Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.431081 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-cdphf"] Dec 01 15:53:30 crc kubenswrapper[4739]: W1201 15:53:30.441078 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod679c7628_a662_4085_8fd9_ae49692f674c.slice/crio-b8aabd7c5137c8226880b11436fb16999a110756cf7a8deee602a48610b0feaf WatchSource:0}: Error finding container b8aabd7c5137c8226880b11436fb16999a110756cf7a8deee602a48610b0feaf: Status 404 returned error can't find the container with id b8aabd7c5137c8226880b11436fb16999a110756cf7a8deee602a48610b0feaf Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.485462 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f38003e3-7781-4602-89e0-0b89e7cf093b" path="/var/lib/kubelet/pods/f38003e3-7781-4602-89e0-0b89e7cf093b/volumes" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.549312 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.652234 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.710858 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-799f5fcdc4-bwhj9"] Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.749903 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-scripts\") pod \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.749996 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-combined-ca-bundle\") pod \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.750025 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-db-sync-config-data\") pod \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.750525 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-config-data\") pod \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.750607 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-etc-machine-id\") pod \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.750661 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8s7v\" (UniqueName: \"kubernetes.io/projected/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-kube-api-access-v8s7v\") pod \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\" (UID: \"588fd2a1-4b85-4af6-8b48-eff7fdf729ba\") " Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.750748 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "588fd2a1-4b85-4af6-8b48-eff7fdf729ba" (UID: "588fd2a1-4b85-4af6-8b48-eff7fdf729ba"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.750965 4739 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.754910 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "588fd2a1-4b85-4af6-8b48-eff7fdf729ba" (UID: "588fd2a1-4b85-4af6-8b48-eff7fdf729ba"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.754936 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-scripts" (OuterVolumeSpecName: "scripts") pod "588fd2a1-4b85-4af6-8b48-eff7fdf729ba" (UID: "588fd2a1-4b85-4af6-8b48-eff7fdf729ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.755285 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-kube-api-access-v8s7v" (OuterVolumeSpecName: "kube-api-access-v8s7v") pod "588fd2a1-4b85-4af6-8b48-eff7fdf729ba" (UID: "588fd2a1-4b85-4af6-8b48-eff7fdf729ba"). InnerVolumeSpecName "kube-api-access-v8s7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.775496 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "588fd2a1-4b85-4af6-8b48-eff7fdf729ba" (UID: "588fd2a1-4b85-4af6-8b48-eff7fdf729ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.796803 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-config-data" (OuterVolumeSpecName: "config-data") pod "588fd2a1-4b85-4af6-8b48-eff7fdf729ba" (UID: "588fd2a1-4b85-4af6-8b48-eff7fdf729ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.852807 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8s7v\" (UniqueName: \"kubernetes.io/projected/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-kube-api-access-v8s7v\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.852861 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.852880 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.852898 4739 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:30 crc kubenswrapper[4739]: I1201 15:53:30.852915 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/588fd2a1-4b85-4af6-8b48-eff7fdf729ba-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.322495 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-g9q4b" event={"ID":"588fd2a1-4b85-4af6-8b48-eff7fdf729ba","Type":"ContainerDied","Data":"5b161810ddf4e767c15b392ce731a74cf2e75d8b50bd1d5a39efec9d91023ad0"} Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.322662 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b161810ddf4e767c15b392ce731a74cf2e75d8b50bd1d5a39efec9d91023ad0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.322547 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-g9q4b" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.335770 4739 generic.go:334] "Generic (PLEG): container finished" podID="679c7628-a662-4085-8fd9-ae49692f674c" containerID="6070603f781536b6853158b705e1c565ea2708e4d926af9c84c24ce08baba464" exitCode=0 Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.335866 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-cdphf" event={"ID":"679c7628-a662-4085-8fd9-ae49692f674c","Type":"ContainerDied","Data":"6070603f781536b6853158b705e1c565ea2708e4d926af9c84c24ce08baba464"} Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.335899 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-cdphf" event={"ID":"679c7628-a662-4085-8fd9-ae49692f674c","Type":"ContainerStarted","Data":"b8aabd7c5137c8226880b11436fb16999a110756cf7a8deee602a48610b0feaf"} Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.343285 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d79c5700-e1a6-416b-bba1-5c211439c9d9","Type":"ContainerStarted","Data":"caf209b4d2ee1eb81d3f05a38cdc7cf0fa764be103faeb2ea353a40594224d76"} Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.343323 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d79c5700-e1a6-416b-bba1-5c211439c9d9","Type":"ContainerStarted","Data":"38bfbd7e528d0885e7d8a70130838f8fb684d4506bfd857ad5ef478a4cd70aca"} Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.367284 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-799f5fcdc4-bwhj9" event={"ID":"3703361d-b3f1-4139-85da-cf21e825b7f2","Type":"ContainerStarted","Data":"cb69ae7e451089101b165f5812a20a66fbcff3a736c206834d1860d6bf3bd0f8"} Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.367331 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-799f5fcdc4-bwhj9" event={"ID":"3703361d-b3f1-4139-85da-cf21e825b7f2","Type":"ContainerStarted","Data":"de3766274b2e61080a4ae14648f3d44fc5582e6ba32fbcfb4278277b14bcdadc"} Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.367346 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-799f5fcdc4-bwhj9" event={"ID":"3703361d-b3f1-4139-85da-cf21e825b7f2","Type":"ContainerStarted","Data":"4c0bce3f28acd172342cc7f86639441c813e5e9ebe707209aeee590004b8c579"} Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.368287 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.368315 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.372178 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" event={"ID":"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd","Type":"ContainerStarted","Data":"9ab03499376e6e5cd4fbd5e02cd987ffc3c8d139fda642da7a974fa2d5075580"} Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.395850 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-799f5fcdc4-bwhj9" podStartSLOduration=2.395828485 podStartE2EDuration="2.395828485s" podCreationTimestamp="2025-12-01 15:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:53:31.386975912 +0000 UTC m=+1113.212722006" watchObservedRunningTime="2025-12-01 15:53:31.395828485 +0000 UTC m=+1113.221574579" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.534060 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 15:53:31 crc kubenswrapper[4739]: E1201 15:53:31.534766 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="588fd2a1-4b85-4af6-8b48-eff7fdf729ba" containerName="cinder-db-sync" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.534777 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="588fd2a1-4b85-4af6-8b48-eff7fdf729ba" containerName="cinder-db-sync" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.534945 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="588fd2a1-4b85-4af6-8b48-eff7fdf729ba" containerName="cinder-db-sync" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.535761 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.538655 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.538837 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.539034 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.539178 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5dk2j" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.553740 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.596121 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-cdphf"] Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.648396 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-rrz9k"] Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.649891 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.667125 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-rrz9k"] Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.668528 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.668593 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-scripts\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.668612 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76ccf696-2126-4fa1-b6ec-1a35c1251e90-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.668633 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-config-data\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.668681 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.668716 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqrn5\" (UniqueName: \"kubernetes.io/projected/76ccf696-2126-4fa1-b6ec-1a35c1251e90-kube-api-access-cqrn5\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.730451 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.731923 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.739746 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.754975 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.770334 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.770410 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-scripts\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.770442 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76ccf696-2126-4fa1-b6ec-1a35c1251e90-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.770460 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-config-data\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.770502 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-rrz9k\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.770521 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.770556 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-rrz9k\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.770575 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqrn5\" (UniqueName: \"kubernetes.io/projected/76ccf696-2126-4fa1-b6ec-1a35c1251e90-kube-api-access-cqrn5\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.770614 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-config\") pod \"dnsmasq-dns-6d97fcdd8f-rrz9k\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.770635 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4db64\" (UniqueName: \"kubernetes.io/projected/ee1eaf87-373f-4138-9625-2be87b6aaa7f-kube-api-access-4db64\") pod \"dnsmasq-dns-6d97fcdd8f-rrz9k\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.770657 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-rrz9k\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.770749 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76ccf696-2126-4fa1-b6ec-1a35c1251e90-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.775341 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-config-data\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.781974 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-scripts\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.782507 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.782888 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.800014 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqrn5\" (UniqueName: \"kubernetes.io/projected/76ccf696-2126-4fa1-b6ec-1a35c1251e90-kube-api-access-cqrn5\") pod \"cinder-scheduler-0\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.872286 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.872348 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3be7d471-1ffd-40c1-bd88-d953b1e06be4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.872396 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-config-data\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.872450 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-rrz9k\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.872478 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsgqh\" (UniqueName: \"kubernetes.io/projected/3be7d471-1ffd-40c1-bd88-d953b1e06be4-kube-api-access-lsgqh\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.872538 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-rrz9k\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.872611 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-config\") pod \"dnsmasq-dns-6d97fcdd8f-rrz9k\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.872633 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3be7d471-1ffd-40c1-bd88-d953b1e06be4-logs\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.872669 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-scripts\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.872687 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4db64\" (UniqueName: \"kubernetes.io/projected/ee1eaf87-373f-4138-9625-2be87b6aaa7f-kube-api-access-4db64\") pod \"dnsmasq-dns-6d97fcdd8f-rrz9k\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.872703 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-config-data-custom\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.872723 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-rrz9k\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.873346 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-config\") pod \"dnsmasq-dns-6d97fcdd8f-rrz9k\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.874831 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.875712 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-dns-svc\") pod \"dnsmasq-dns-6d97fcdd8f-rrz9k\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.876799 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-ovsdbserver-sb\") pod \"dnsmasq-dns-6d97fcdd8f-rrz9k\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.877071 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-ovsdbserver-nb\") pod \"dnsmasq-dns-6d97fcdd8f-rrz9k\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.895994 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4db64\" (UniqueName: \"kubernetes.io/projected/ee1eaf87-373f-4138-9625-2be87b6aaa7f-kube-api-access-4db64\") pod \"dnsmasq-dns-6d97fcdd8f-rrz9k\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.973456 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.974675 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-config-data\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.974729 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsgqh\" (UniqueName: \"kubernetes.io/projected/3be7d471-1ffd-40c1-bd88-d953b1e06be4-kube-api-access-lsgqh\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.974796 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3be7d471-1ffd-40c1-bd88-d953b1e06be4-logs\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.974824 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-scripts\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.974845 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-config-data-custom\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.974885 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.974926 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3be7d471-1ffd-40c1-bd88-d953b1e06be4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.975019 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3be7d471-1ffd-40c1-bd88-d953b1e06be4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.975979 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3be7d471-1ffd-40c1-bd88-d953b1e06be4-logs\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.979704 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-config-data-custom\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.981227 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-scripts\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.983924 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.990384 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsgqh\" (UniqueName: \"kubernetes.io/projected/3be7d471-1ffd-40c1-bd88-d953b1e06be4-kube-api-access-lsgqh\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:31 crc kubenswrapper[4739]: I1201 15:53:31.990455 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-config-data\") pod \"cinder-api-0\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " pod="openstack/cinder-api-0" Dec 01 15:53:32 crc kubenswrapper[4739]: I1201 15:53:32.057738 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 15:53:32 crc kubenswrapper[4739]: I1201 15:53:32.781150 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 15:53:32 crc kubenswrapper[4739]: I1201 15:53:32.902278 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-rrz9k"] Dec 01 15:53:32 crc kubenswrapper[4739]: W1201 15:53:32.913303 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee1eaf87_373f_4138_9625_2be87b6aaa7f.slice/crio-79c61688269609b18bda7f09a8b479ccaf967fbbe82dc8c2e70ec0c82c7184bb WatchSource:0}: Error finding container 79c61688269609b18bda7f09a8b479ccaf967fbbe82dc8c2e70ec0c82c7184bb: Status 404 returned error can't find the container with id 79c61688269609b18bda7f09a8b479ccaf967fbbe82dc8c2e70ec0c82c7184bb Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.030092 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.405753 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"76ccf696-2126-4fa1-b6ec-1a35c1251e90","Type":"ContainerStarted","Data":"7edab46af5fae2b786a1e90baf91a7bf63a22ec5ac252d8cc9859671304cb8cf"} Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.409490 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" event={"ID":"02f9d262-1e44-42e4-a413-b85bb5f25241","Type":"ContainerStarted","Data":"331ab8fa5ba54848f42be81379d236888d2269238293d1eac8f6b58e3fe89d8b"} Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.409516 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" event={"ID":"02f9d262-1e44-42e4-a413-b85bb5f25241","Type":"ContainerStarted","Data":"0b5aba87b3ff7d7e194445dc1e96986a5f3b7d7d195095d2a351a9aca520f082"} Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.412434 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3be7d471-1ffd-40c1-bd88-d953b1e06be4","Type":"ContainerStarted","Data":"7eba162cce1d186be3492d07db471443fcb2aea16aa9cafcab82becd404952bb"} Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.414165 4739 generic.go:334] "Generic (PLEG): container finished" podID="ee1eaf87-373f-4138-9625-2be87b6aaa7f" containerID="397e83467e0d2793e2cdb6411b8ef238990c4bb95f28f6f23ca00cb4953c3457" exitCode=0 Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.414258 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" event={"ID":"ee1eaf87-373f-4138-9625-2be87b6aaa7f","Type":"ContainerDied","Data":"397e83467e0d2793e2cdb6411b8ef238990c4bb95f28f6f23ca00cb4953c3457"} Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.414304 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" event={"ID":"ee1eaf87-373f-4138-9625-2be87b6aaa7f","Type":"ContainerStarted","Data":"79c61688269609b18bda7f09a8b479ccaf967fbbe82dc8c2e70ec0c82c7184bb"} Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.417691 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb684768f-cdphf" podUID="679c7628-a662-4085-8fd9-ae49692f674c" containerName="dnsmasq-dns" containerID="cri-o://1c283989e58538280ebafa29677a8c53b8ce1ccfa6d7dd8b4e307bbb9cd5aa18" gracePeriod=10 Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.417757 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-cdphf" event={"ID":"679c7628-a662-4085-8fd9-ae49692f674c","Type":"ContainerStarted","Data":"1c283989e58538280ebafa29677a8c53b8ce1ccfa6d7dd8b4e307bbb9cd5aa18"} Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.417787 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.426330 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d79c5700-e1a6-416b-bba1-5c211439c9d9","Type":"ContainerStarted","Data":"d34f390505dfaf24ddeea1b458f02ef3eb977f5d31823a8f72cd3ef73d02b41f"} Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.434952 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6bbd6b9f5c-wjbz6" podStartSLOduration=2.315472153 podStartE2EDuration="4.434936659s" podCreationTimestamp="2025-12-01 15:53:29 +0000 UTC" firstStartedPulling="2025-12-01 15:53:30.221632801 +0000 UTC m=+1112.047378895" lastFinishedPulling="2025-12-01 15:53:32.341097307 +0000 UTC m=+1114.166843401" observedRunningTime="2025-12-01 15:53:33.43206649 +0000 UTC m=+1115.257812584" watchObservedRunningTime="2025-12-01 15:53:33.434936659 +0000 UTC m=+1115.260682743" Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.440646 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" event={"ID":"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd","Type":"ContainerStarted","Data":"f390f4baa05798e17d682c4036b0a2cf9c68171e7ef904d158c5be5e3758eec6"} Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.440677 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" event={"ID":"2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd","Type":"ContainerStarted","Data":"a627891a738c8438b36a59d1c30cc4476c561a68c2f06019428cbb5301413357"} Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.484975 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb684768f-cdphf" podStartSLOduration=4.484954544 podStartE2EDuration="4.484954544s" podCreationTimestamp="2025-12-01 15:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:53:33.460777126 +0000 UTC m=+1115.286523240" watchObservedRunningTime="2025-12-01 15:53:33.484954544 +0000 UTC m=+1115.310700638" Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.513806 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-ff455997d-9lg5c" podStartSLOduration=2.500112877 podStartE2EDuration="4.513786234s" podCreationTimestamp="2025-12-01 15:53:29 +0000 UTC" firstStartedPulling="2025-12-01 15:53:30.326986296 +0000 UTC m=+1112.152732390" lastFinishedPulling="2025-12-01 15:53:32.340659653 +0000 UTC m=+1114.166405747" observedRunningTime="2025-12-01 15:53:33.493868259 +0000 UTC m=+1115.319614353" watchObservedRunningTime="2025-12-01 15:53:33.513786234 +0000 UTC m=+1115.339532328" Dec 01 15:53:33 crc kubenswrapper[4739]: I1201 15:53:33.990578 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.126855 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-ovsdbserver-sb\") pod \"679c7628-a662-4085-8fd9-ae49692f674c\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.127006 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8gzt\" (UniqueName: \"kubernetes.io/projected/679c7628-a662-4085-8fd9-ae49692f674c-kube-api-access-k8gzt\") pod \"679c7628-a662-4085-8fd9-ae49692f674c\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.127041 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-ovsdbserver-nb\") pod \"679c7628-a662-4085-8fd9-ae49692f674c\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.127087 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-config\") pod \"679c7628-a662-4085-8fd9-ae49692f674c\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.127118 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-dns-svc\") pod \"679c7628-a662-4085-8fd9-ae49692f674c\" (UID: \"679c7628-a662-4085-8fd9-ae49692f674c\") " Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.133178 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/679c7628-a662-4085-8fd9-ae49692f674c-kube-api-access-k8gzt" (OuterVolumeSpecName: "kube-api-access-k8gzt") pod "679c7628-a662-4085-8fd9-ae49692f674c" (UID: "679c7628-a662-4085-8fd9-ae49692f674c"). InnerVolumeSpecName "kube-api-access-k8gzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.194281 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-config" (OuterVolumeSpecName: "config") pod "679c7628-a662-4085-8fd9-ae49692f674c" (UID: "679c7628-a662-4085-8fd9-ae49692f674c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.197094 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "679c7628-a662-4085-8fd9-ae49692f674c" (UID: "679c7628-a662-4085-8fd9-ae49692f674c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.201011 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "679c7628-a662-4085-8fd9-ae49692f674c" (UID: "679c7628-a662-4085-8fd9-ae49692f674c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.230068 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.230098 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8gzt\" (UniqueName: \"kubernetes.io/projected/679c7628-a662-4085-8fd9-ae49692f674c-kube-api-access-k8gzt\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.230189 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.230198 4739 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.236825 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "679c7628-a662-4085-8fd9-ae49692f674c" (UID: "679c7628-a662-4085-8fd9-ae49692f674c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.331404 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/679c7628-a662-4085-8fd9-ae49692f674c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.452335 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d79c5700-e1a6-416b-bba1-5c211439c9d9","Type":"ContainerStarted","Data":"d3bb11e2c8f408e8499881b7bb9857a7ad272ecb71d08fcc893c02aeb5afb5b4"} Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.454308 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3be7d471-1ffd-40c1-bd88-d953b1e06be4","Type":"ContainerStarted","Data":"204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986"} Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.455818 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" event={"ID":"ee1eaf87-373f-4138-9625-2be87b6aaa7f","Type":"ContainerStarted","Data":"16e6b948b38abbcd45d6da8fdbb50b5b859f6584246db6e9c48398ba169b07e4"} Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.456895 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.459001 4739 generic.go:334] "Generic (PLEG): container finished" podID="679c7628-a662-4085-8fd9-ae49692f674c" containerID="1c283989e58538280ebafa29677a8c53b8ce1ccfa6d7dd8b4e307bbb9cd5aa18" exitCode=0 Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.459061 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-cdphf" event={"ID":"679c7628-a662-4085-8fd9-ae49692f674c","Type":"ContainerDied","Data":"1c283989e58538280ebafa29677a8c53b8ce1ccfa6d7dd8b4e307bbb9cd5aa18"} Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.459077 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb684768f-cdphf" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.459252 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb684768f-cdphf" event={"ID":"679c7628-a662-4085-8fd9-ae49692f674c","Type":"ContainerDied","Data":"b8aabd7c5137c8226880b11436fb16999a110756cf7a8deee602a48610b0feaf"} Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.459359 4739 scope.go:117] "RemoveContainer" containerID="1c283989e58538280ebafa29677a8c53b8ce1ccfa6d7dd8b4e307bbb9cd5aa18" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.474067 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" podStartSLOduration=3.4740527390000002 podStartE2EDuration="3.474052739s" podCreationTimestamp="2025-12-01 15:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:53:34.471895562 +0000 UTC m=+1116.297641656" watchObservedRunningTime="2025-12-01 15:53:34.474052739 +0000 UTC m=+1116.299798833" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.518902 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-cdphf"] Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.529946 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb684768f-cdphf"] Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.530537 4739 scope.go:117] "RemoveContainer" containerID="6070603f781536b6853158b705e1c565ea2708e4d926af9c84c24ce08baba464" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.630725 4739 scope.go:117] "RemoveContainer" containerID="1c283989e58538280ebafa29677a8c53b8ce1ccfa6d7dd8b4e307bbb9cd5aa18" Dec 01 15:53:34 crc kubenswrapper[4739]: E1201 15:53:34.632660 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c283989e58538280ebafa29677a8c53b8ce1ccfa6d7dd8b4e307bbb9cd5aa18\": container with ID starting with 1c283989e58538280ebafa29677a8c53b8ce1ccfa6d7dd8b4e307bbb9cd5aa18 not found: ID does not exist" containerID="1c283989e58538280ebafa29677a8c53b8ce1ccfa6d7dd8b4e307bbb9cd5aa18" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.632702 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c283989e58538280ebafa29677a8c53b8ce1ccfa6d7dd8b4e307bbb9cd5aa18"} err="failed to get container status \"1c283989e58538280ebafa29677a8c53b8ce1ccfa6d7dd8b4e307bbb9cd5aa18\": rpc error: code = NotFound desc = could not find container \"1c283989e58538280ebafa29677a8c53b8ce1ccfa6d7dd8b4e307bbb9cd5aa18\": container with ID starting with 1c283989e58538280ebafa29677a8c53b8ce1ccfa6d7dd8b4e307bbb9cd5aa18 not found: ID does not exist" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.632730 4739 scope.go:117] "RemoveContainer" containerID="6070603f781536b6853158b705e1c565ea2708e4d926af9c84c24ce08baba464" Dec 01 15:53:34 crc kubenswrapper[4739]: E1201 15:53:34.633195 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6070603f781536b6853158b705e1c565ea2708e4d926af9c84c24ce08baba464\": container with ID starting with 6070603f781536b6853158b705e1c565ea2708e4d926af9c84c24ce08baba464 not found: ID does not exist" containerID="6070603f781536b6853158b705e1c565ea2708e4d926af9c84c24ce08baba464" Dec 01 15:53:34 crc kubenswrapper[4739]: I1201 15:53:34.633219 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6070603f781536b6853158b705e1c565ea2708e4d926af9c84c24ce08baba464"} err="failed to get container status \"6070603f781536b6853158b705e1c565ea2708e4d926af9c84c24ce08baba464\": rpc error: code = NotFound desc = could not find container \"6070603f781536b6853158b705e1c565ea2708e4d926af9c84c24ce08baba464\": container with ID starting with 6070603f781536b6853158b705e1c565ea2708e4d926af9c84c24ce08baba464 not found: ID does not exist" Dec 01 15:53:35 crc kubenswrapper[4739]: I1201 15:53:35.490899 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"76ccf696-2126-4fa1-b6ec-1a35c1251e90","Type":"ContainerStarted","Data":"0963a9c722c8dcb52061f80b83ed810edc4e8229ab029f83e7bec46983a3ed5f"} Dec 01 15:53:35 crc kubenswrapper[4739]: I1201 15:53:35.502737 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3be7d471-1ffd-40c1-bd88-d953b1e06be4","Type":"ContainerStarted","Data":"67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379"} Dec 01 15:53:35 crc kubenswrapper[4739]: I1201 15:53:35.502860 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 15:53:35 crc kubenswrapper[4739]: I1201 15:53:35.811893 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.811866098 podStartE2EDuration="4.811866098s" podCreationTimestamp="2025-12-01 15:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:53:35.534587502 +0000 UTC m=+1117.360333606" watchObservedRunningTime="2025-12-01 15:53:35.811866098 +0000 UTC m=+1117.637612202" Dec 01 15:53:35 crc kubenswrapper[4739]: I1201 15:53:35.814983 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.227714 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-856bfcfff6-5wtcc"] Dec 01 15:53:36 crc kubenswrapper[4739]: E1201 15:53:36.228365 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="679c7628-a662-4085-8fd9-ae49692f674c" containerName="dnsmasq-dns" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.228382 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="679c7628-a662-4085-8fd9-ae49692f674c" containerName="dnsmasq-dns" Dec 01 15:53:36 crc kubenswrapper[4739]: E1201 15:53:36.228407 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="679c7628-a662-4085-8fd9-ae49692f674c" containerName="init" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.228427 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="679c7628-a662-4085-8fd9-ae49692f674c" containerName="init" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.228590 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="679c7628-a662-4085-8fd9-ae49692f674c" containerName="dnsmasq-dns" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.229498 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.234328 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.236707 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.244708 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-856bfcfff6-5wtcc"] Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.370960 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-config-data-custom\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.371389 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-222fz\" (UniqueName: \"kubernetes.io/projected/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-kube-api-access-222fz\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.371466 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-public-tls-certs\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.371546 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-logs\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.371569 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-config-data\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.371601 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-internal-tls-certs\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.371659 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-combined-ca-bundle\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.477147 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-222fz\" (UniqueName: \"kubernetes.io/projected/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-kube-api-access-222fz\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.477242 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-public-tls-certs\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.477347 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-logs\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.477383 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-config-data\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.477437 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-internal-tls-certs\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.477513 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-combined-ca-bundle\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.477584 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-config-data-custom\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.478957 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-logs\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.484795 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-combined-ca-bundle\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.490512 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-config-data-custom\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.494669 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-public-tls-certs\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.497340 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-internal-tls-certs\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.498019 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-222fz\" (UniqueName: \"kubernetes.io/projected/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-kube-api-access-222fz\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.498030 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff-config-data\") pod \"barbican-api-856bfcfff6-5wtcc\" (UID: \"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff\") " pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.512246 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="679c7628-a662-4085-8fd9-ae49692f674c" path="/var/lib/kubelet/pods/679c7628-a662-4085-8fd9-ae49692f674c/volumes" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.537884 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"76ccf696-2126-4fa1-b6ec-1a35c1251e90","Type":"ContainerStarted","Data":"12e3e6f9a0de70e3cb685e1c27e2d3b556e71dd38a6c45ad5b4d57ae9e923208"} Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.540951 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d79c5700-e1a6-416b-bba1-5c211439c9d9","Type":"ContainerStarted","Data":"3d096b0f728cd7c50a87d1e77a88cd58239e705a29aaa6fa8ee016ed1cfdb9a9"} Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.541313 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.568016 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.739626876 podStartE2EDuration="5.567998816s" podCreationTimestamp="2025-12-01 15:53:31 +0000 UTC" firstStartedPulling="2025-12-01 15:53:32.797239868 +0000 UTC m=+1114.622985962" lastFinishedPulling="2025-12-01 15:53:33.625611808 +0000 UTC m=+1115.451357902" observedRunningTime="2025-12-01 15:53:36.553165539 +0000 UTC m=+1118.378911633" watchObservedRunningTime="2025-12-01 15:53:36.567998816 +0000 UTC m=+1118.393744900" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.593838 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.845990 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.868823 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.335311625 podStartE2EDuration="7.868800389s" podCreationTimestamp="2025-12-01 15:53:29 +0000 UTC" firstStartedPulling="2025-12-01 15:53:30.56891913 +0000 UTC m=+1112.394665224" lastFinishedPulling="2025-12-01 15:53:36.102407894 +0000 UTC m=+1117.928153988" observedRunningTime="2025-12-01 15:53:36.583731853 +0000 UTC m=+1118.409477947" watchObservedRunningTime="2025-12-01 15:53:36.868800389 +0000 UTC m=+1118.694546493" Dec 01 15:53:36 crc kubenswrapper[4739]: I1201 15:53:36.876516 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 15:53:37 crc kubenswrapper[4739]: I1201 15:53:37.064354 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-856bfcfff6-5wtcc"] Dec 01 15:53:37 crc kubenswrapper[4739]: W1201 15:53:37.072193 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb34d197e_14ff_4c2e_b1e2_a6fde0dde4ff.slice/crio-6fde19553e2f9c5f4fc345500543af2c1ffb1eb0f7d0c47fe751234c82f661a8 WatchSource:0}: Error finding container 6fde19553e2f9c5f4fc345500543af2c1ffb1eb0f7d0c47fe751234c82f661a8: Status 404 returned error can't find the container with id 6fde19553e2f9c5f4fc345500543af2c1ffb1eb0f7d0c47fe751234c82f661a8 Dec 01 15:53:37 crc kubenswrapper[4739]: I1201 15:53:37.553720 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-856bfcfff6-5wtcc" event={"ID":"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff","Type":"ContainerStarted","Data":"3bc4b1e52d5b78aad9cada5c0f7d8ffefab38dc82e47e4ef8be225da935bbfe9"} Dec 01 15:53:37 crc kubenswrapper[4739]: I1201 15:53:37.555010 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-856bfcfff6-5wtcc" event={"ID":"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff","Type":"ContainerStarted","Data":"6fde19553e2f9c5f4fc345500543af2c1ffb1eb0f7d0c47fe751234c82f661a8"} Dec 01 15:53:37 crc kubenswrapper[4739]: I1201 15:53:37.555833 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3be7d471-1ffd-40c1-bd88-d953b1e06be4" containerName="cinder-api-log" containerID="cri-o://204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986" gracePeriod=30 Dec 01 15:53:37 crc kubenswrapper[4739]: I1201 15:53:37.556345 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3be7d471-1ffd-40c1-bd88-d953b1e06be4" containerName="cinder-api" containerID="cri-o://67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379" gracePeriod=30 Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.231204 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.320775 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-config-data-custom\") pod \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.320886 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsgqh\" (UniqueName: \"kubernetes.io/projected/3be7d471-1ffd-40c1-bd88-d953b1e06be4-kube-api-access-lsgqh\") pod \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.320916 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3be7d471-1ffd-40c1-bd88-d953b1e06be4-etc-machine-id\") pod \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.320971 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-scripts\") pod \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.321057 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-combined-ca-bundle\") pod \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.321097 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3be7d471-1ffd-40c1-bd88-d953b1e06be4-logs\") pod \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.321129 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-config-data\") pod \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\" (UID: \"3be7d471-1ffd-40c1-bd88-d953b1e06be4\") " Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.322391 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3be7d471-1ffd-40c1-bd88-d953b1e06be4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3be7d471-1ffd-40c1-bd88-d953b1e06be4" (UID: "3be7d471-1ffd-40c1-bd88-d953b1e06be4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.327101 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3be7d471-1ffd-40c1-bd88-d953b1e06be4" (UID: "3be7d471-1ffd-40c1-bd88-d953b1e06be4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.332831 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3be7d471-1ffd-40c1-bd88-d953b1e06be4-logs" (OuterVolumeSpecName: "logs") pod "3be7d471-1ffd-40c1-bd88-d953b1e06be4" (UID: "3be7d471-1ffd-40c1-bd88-d953b1e06be4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.334876 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-scripts" (OuterVolumeSpecName: "scripts") pod "3be7d471-1ffd-40c1-bd88-d953b1e06be4" (UID: "3be7d471-1ffd-40c1-bd88-d953b1e06be4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.348809 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3be7d471-1ffd-40c1-bd88-d953b1e06be4-kube-api-access-lsgqh" (OuterVolumeSpecName: "kube-api-access-lsgqh") pod "3be7d471-1ffd-40c1-bd88-d953b1e06be4" (UID: "3be7d471-1ffd-40c1-bd88-d953b1e06be4"). InnerVolumeSpecName "kube-api-access-lsgqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.361804 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3be7d471-1ffd-40c1-bd88-d953b1e06be4" (UID: "3be7d471-1ffd-40c1-bd88-d953b1e06be4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.383485 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-config-data" (OuterVolumeSpecName: "config-data") pod "3be7d471-1ffd-40c1-bd88-d953b1e06be4" (UID: "3be7d471-1ffd-40c1-bd88-d953b1e06be4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.423410 4739 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3be7d471-1ffd-40c1-bd88-d953b1e06be4-logs\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.423462 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.423474 4739 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.423489 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsgqh\" (UniqueName: \"kubernetes.io/projected/3be7d471-1ffd-40c1-bd88-d953b1e06be4-kube-api-access-lsgqh\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.423501 4739 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3be7d471-1ffd-40c1-bd88-d953b1e06be4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.423512 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.423521 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be7d471-1ffd-40c1-bd88-d953b1e06be4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.561230 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-856bfcfff6-5wtcc" event={"ID":"b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff","Type":"ContainerStarted","Data":"5a30a141046750e8cdb6ffe58ca9fa9f91bdfbb5346ec5879f75e95dfea87b74"} Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.561343 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.563013 4739 generic.go:334] "Generic (PLEG): container finished" podID="3be7d471-1ffd-40c1-bd88-d953b1e06be4" containerID="67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379" exitCode=0 Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.563037 4739 generic.go:334] "Generic (PLEG): container finished" podID="3be7d471-1ffd-40c1-bd88-d953b1e06be4" containerID="204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986" exitCode=143 Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.563050 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3be7d471-1ffd-40c1-bd88-d953b1e06be4","Type":"ContainerDied","Data":"67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379"} Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.563105 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3be7d471-1ffd-40c1-bd88-d953b1e06be4","Type":"ContainerDied","Data":"204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986"} Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.563121 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3be7d471-1ffd-40c1-bd88-d953b1e06be4","Type":"ContainerDied","Data":"7eba162cce1d186be3492d07db471443fcb2aea16aa9cafcab82becd404952bb"} Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.563068 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.563158 4739 scope.go:117] "RemoveContainer" containerID="67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.606084 4739 scope.go:117] "RemoveContainer" containerID="204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.621320 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-856bfcfff6-5wtcc" podStartSLOduration=2.621304089 podStartE2EDuration="2.621304089s" podCreationTimestamp="2025-12-01 15:53:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:53:38.59384345 +0000 UTC m=+1120.419589534" watchObservedRunningTime="2025-12-01 15:53:38.621304089 +0000 UTC m=+1120.447050183" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.631747 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.631924 4739 scope.go:117] "RemoveContainer" containerID="67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379" Dec 01 15:53:38 crc kubenswrapper[4739]: E1201 15:53:38.632370 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379\": container with ID starting with 67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379 not found: ID does not exist" containerID="67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.632402 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379"} err="failed to get container status \"67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379\": rpc error: code = NotFound desc = could not find container \"67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379\": container with ID starting with 67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379 not found: ID does not exist" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.632438 4739 scope.go:117] "RemoveContainer" containerID="204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986" Dec 01 15:53:38 crc kubenswrapper[4739]: E1201 15:53:38.633246 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986\": container with ID starting with 204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986 not found: ID does not exist" containerID="204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.633266 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986"} err="failed to get container status \"204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986\": rpc error: code = NotFound desc = could not find container \"204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986\": container with ID starting with 204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986 not found: ID does not exist" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.633279 4739 scope.go:117] "RemoveContainer" containerID="67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.633578 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379"} err="failed to get container status \"67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379\": rpc error: code = NotFound desc = could not find container \"67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379\": container with ID starting with 67a8c8a859f9a6152d13c3eb3ec3040fd6872314002ca6a52394b92b5f35f379 not found: ID does not exist" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.633609 4739 scope.go:117] "RemoveContainer" containerID="204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.633811 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986"} err="failed to get container status \"204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986\": rpc error: code = NotFound desc = could not find container \"204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986\": container with ID starting with 204a18ae68ddc841fb0cf9f1e42efbb618b950b09c0329366728eb7c7177b986 not found: ID does not exist" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.638924 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.650835 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.666231 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 15:53:38 crc kubenswrapper[4739]: E1201 15:53:38.666636 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3be7d471-1ffd-40c1-bd88-d953b1e06be4" containerName="cinder-api-log" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.666653 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="3be7d471-1ffd-40c1-bd88-d953b1e06be4" containerName="cinder-api-log" Dec 01 15:53:38 crc kubenswrapper[4739]: E1201 15:53:38.666748 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3be7d471-1ffd-40c1-bd88-d953b1e06be4" containerName="cinder-api" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.666757 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="3be7d471-1ffd-40c1-bd88-d953b1e06be4" containerName="cinder-api" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.666927 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="3be7d471-1ffd-40c1-bd88-d953b1e06be4" containerName="cinder-api" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.666951 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="3be7d471-1ffd-40c1-bd88-d953b1e06be4" containerName="cinder-api-log" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.667990 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.670063 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.670682 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.670927 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.680169 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.831977 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlb9v\" (UniqueName: \"kubernetes.io/projected/32d670bc-bba0-45ed-80c9-f4258d97452c-kube-api-access-jlb9v\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.832037 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-config-data\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.832099 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-scripts\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.832126 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32d670bc-bba0-45ed-80c9-f4258d97452c-logs\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.832167 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.832187 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/32d670bc-bba0-45ed-80c9-f4258d97452c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.832279 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.832409 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-config-data-custom\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.832483 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.933716 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-scripts\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.933777 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32d670bc-bba0-45ed-80c9-f4258d97452c-logs\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.933830 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.933861 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/32d670bc-bba0-45ed-80c9-f4258d97452c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.933891 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.933936 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-config-data-custom\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.933966 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.933982 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/32d670bc-bba0-45ed-80c9-f4258d97452c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.934398 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32d670bc-bba0-45ed-80c9-f4258d97452c-logs\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.934062 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlb9v\" (UniqueName: \"kubernetes.io/projected/32d670bc-bba0-45ed-80c9-f4258d97452c-kube-api-access-jlb9v\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.934848 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-config-data\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.938038 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-scripts\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.938954 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-config-data-custom\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.939375 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-config-data\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.940679 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.944557 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.951271 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32d670bc-bba0-45ed-80c9-f4258d97452c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.954335 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlb9v\" (UniqueName: \"kubernetes.io/projected/32d670bc-bba0-45ed-80c9-f4258d97452c-kube-api-access-jlb9v\") pod \"cinder-api-0\" (UID: \"32d670bc-bba0-45ed-80c9-f4258d97452c\") " pod="openstack/cinder-api-0" Dec 01 15:53:38 crc kubenswrapper[4739]: I1201 15:53:38.985536 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 15:53:39 crc kubenswrapper[4739]: I1201 15:53:39.479477 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 15:53:39 crc kubenswrapper[4739]: W1201 15:53:39.480892 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32d670bc_bba0_45ed_80c9_f4258d97452c.slice/crio-ece1f0b51fbed3f066b136e4ef85ecf5fb99cc3f8f1aa68586186c14ed1c2ef9 WatchSource:0}: Error finding container ece1f0b51fbed3f066b136e4ef85ecf5fb99cc3f8f1aa68586186c14ed1c2ef9: Status 404 returned error can't find the container with id ece1f0b51fbed3f066b136e4ef85ecf5fb99cc3f8f1aa68586186c14ed1c2ef9 Dec 01 15:53:39 crc kubenswrapper[4739]: I1201 15:53:39.574086 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"32d670bc-bba0-45ed-80c9-f4258d97452c","Type":"ContainerStarted","Data":"ece1f0b51fbed3f066b136e4ef85ecf5fb99cc3f8f1aa68586186c14ed1c2ef9"} Dec 01 15:53:39 crc kubenswrapper[4739]: I1201 15:53:39.574141 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:39 crc kubenswrapper[4739]: I1201 15:53:39.622567 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:53:39 crc kubenswrapper[4739]: I1201 15:53:39.622625 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:53:39 crc kubenswrapper[4739]: I1201 15:53:39.622673 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:53:39 crc kubenswrapper[4739]: I1201 15:53:39.623475 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d23f235659cb740130c9b9ee149b02d7de5733c24075029cc9884583e5415dad"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:53:39 crc kubenswrapper[4739]: I1201 15:53:39.623545 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://d23f235659cb740130c9b9ee149b02d7de5733c24075029cc9884583e5415dad" gracePeriod=600 Dec 01 15:53:40 crc kubenswrapper[4739]: I1201 15:53:40.449834 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:40 crc kubenswrapper[4739]: I1201 15:53:40.491659 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3be7d471-1ffd-40c1-bd88-d953b1e06be4" path="/var/lib/kubelet/pods/3be7d471-1ffd-40c1-bd88-d953b1e06be4/volumes" Dec 01 15:53:40 crc kubenswrapper[4739]: I1201 15:53:40.586290 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="d23f235659cb740130c9b9ee149b02d7de5733c24075029cc9884583e5415dad" exitCode=0 Dec 01 15:53:40 crc kubenswrapper[4739]: I1201 15:53:40.586354 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"d23f235659cb740130c9b9ee149b02d7de5733c24075029cc9884583e5415dad"} Dec 01 15:53:40 crc kubenswrapper[4739]: I1201 15:53:40.586385 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"faad8f9bb664920c3ea11c12b9b5f870015e0ec81598417117215899a38ccfdb"} Dec 01 15:53:40 crc kubenswrapper[4739]: I1201 15:53:40.586401 4739 scope.go:117] "RemoveContainer" containerID="c739ec8ed0d32f48e79459013f2a8032f453df5dc42b21acdeda98c2b463c62e" Dec 01 15:53:40 crc kubenswrapper[4739]: I1201 15:53:40.594457 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"32d670bc-bba0-45ed-80c9-f4258d97452c","Type":"ContainerStarted","Data":"b5697e6a4d20a75f7c5f109e13f36ed2c74ebf1344c8d8842ca4dd3536090fce"} Dec 01 15:53:41 crc kubenswrapper[4739]: I1201 15:53:41.609824 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"32d670bc-bba0-45ed-80c9-f4258d97452c","Type":"ContainerStarted","Data":"f7acb8dc4d03f2ecf799f8316e257fedcea66ef65e02346d5601afc690c79bd3"} Dec 01 15:53:41 crc kubenswrapper[4739]: I1201 15:53:41.610280 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 15:53:41 crc kubenswrapper[4739]: I1201 15:53:41.649996 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.649970862 podStartE2EDuration="3.649970862s" podCreationTimestamp="2025-12-01 15:53:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:53:41.639136437 +0000 UTC m=+1123.464882591" watchObservedRunningTime="2025-12-01 15:53:41.649970862 +0000 UTC m=+1123.475716996" Dec 01 15:53:41 crc kubenswrapper[4739]: I1201 15:53:41.976670 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.047931 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-vzh4g"] Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.048543 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" podUID="9ce6139b-7b9f-449e-a642-ca71b4740873" containerName="dnsmasq-dns" containerID="cri-o://e4e0e96b248781e6595254f727c731b71c43924c072648704ec9551a9e77679f" gracePeriod=10 Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.166235 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.219374 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.599348 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.619233 4739 generic.go:334] "Generic (PLEG): container finished" podID="9ce6139b-7b9f-449e-a642-ca71b4740873" containerID="e4e0e96b248781e6595254f727c731b71c43924c072648704ec9551a9e77679f" exitCode=0 Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.619532 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="76ccf696-2126-4fa1-b6ec-1a35c1251e90" containerName="cinder-scheduler" containerID="cri-o://0963a9c722c8dcb52061f80b83ed810edc4e8229ab029f83e7bec46983a3ed5f" gracePeriod=30 Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.619829 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.619893 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" event={"ID":"9ce6139b-7b9f-449e-a642-ca71b4740873","Type":"ContainerDied","Data":"e4e0e96b248781e6595254f727c731b71c43924c072648704ec9551a9e77679f"} Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.619959 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b946d459c-vzh4g" event={"ID":"9ce6139b-7b9f-449e-a642-ca71b4740873","Type":"ContainerDied","Data":"61c498a63f17b128ac79c5ee5b1ae2dcb57527222698cc836f1e69beb3e75ad0"} Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.619989 4739 scope.go:117] "RemoveContainer" containerID="e4e0e96b248781e6595254f727c731b71c43924c072648704ec9551a9e77679f" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.620170 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="76ccf696-2126-4fa1-b6ec-1a35c1251e90" containerName="probe" containerID="cri-o://12e3e6f9a0de70e3cb685e1c27e2d3b556e71dd38a6c45ad5b4d57ae9e923208" gracePeriod=30 Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.669831 4739 scope.go:117] "RemoveContainer" containerID="1d457b66af62af0d8203ec52864649b703d76cdf33f558e4d537cb02768b4bf0" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.692745 4739 scope.go:117] "RemoveContainer" containerID="e4e0e96b248781e6595254f727c731b71c43924c072648704ec9551a9e77679f" Dec 01 15:53:42 crc kubenswrapper[4739]: E1201 15:53:42.693348 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4e0e96b248781e6595254f727c731b71c43924c072648704ec9551a9e77679f\": container with ID starting with e4e0e96b248781e6595254f727c731b71c43924c072648704ec9551a9e77679f not found: ID does not exist" containerID="e4e0e96b248781e6595254f727c731b71c43924c072648704ec9551a9e77679f" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.693520 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4e0e96b248781e6595254f727c731b71c43924c072648704ec9551a9e77679f"} err="failed to get container status \"e4e0e96b248781e6595254f727c731b71c43924c072648704ec9551a9e77679f\": rpc error: code = NotFound desc = could not find container \"e4e0e96b248781e6595254f727c731b71c43924c072648704ec9551a9e77679f\": container with ID starting with e4e0e96b248781e6595254f727c731b71c43924c072648704ec9551a9e77679f not found: ID does not exist" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.693638 4739 scope.go:117] "RemoveContainer" containerID="1d457b66af62af0d8203ec52864649b703d76cdf33f558e4d537cb02768b4bf0" Dec 01 15:53:42 crc kubenswrapper[4739]: E1201 15:53:42.694523 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d457b66af62af0d8203ec52864649b703d76cdf33f558e4d537cb02768b4bf0\": container with ID starting with 1d457b66af62af0d8203ec52864649b703d76cdf33f558e4d537cb02768b4bf0 not found: ID does not exist" containerID="1d457b66af62af0d8203ec52864649b703d76cdf33f558e4d537cb02768b4bf0" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.694567 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d457b66af62af0d8203ec52864649b703d76cdf33f558e4d537cb02768b4bf0"} err="failed to get container status \"1d457b66af62af0d8203ec52864649b703d76cdf33f558e4d537cb02768b4bf0\": rpc error: code = NotFound desc = could not find container \"1d457b66af62af0d8203ec52864649b703d76cdf33f558e4d537cb02768b4bf0\": container with ID starting with 1d457b66af62af0d8203ec52864649b703d76cdf33f558e4d537cb02768b4bf0 not found: ID does not exist" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.714794 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-config\") pod \"9ce6139b-7b9f-449e-a642-ca71b4740873\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.715848 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-ovsdbserver-sb\") pod \"9ce6139b-7b9f-449e-a642-ca71b4740873\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.716071 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-dns-svc\") pod \"9ce6139b-7b9f-449e-a642-ca71b4740873\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.716278 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8kxb\" (UniqueName: \"kubernetes.io/projected/9ce6139b-7b9f-449e-a642-ca71b4740873-kube-api-access-l8kxb\") pod \"9ce6139b-7b9f-449e-a642-ca71b4740873\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.716447 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-ovsdbserver-nb\") pod \"9ce6139b-7b9f-449e-a642-ca71b4740873\" (UID: \"9ce6139b-7b9f-449e-a642-ca71b4740873\") " Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.721855 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ce6139b-7b9f-449e-a642-ca71b4740873-kube-api-access-l8kxb" (OuterVolumeSpecName: "kube-api-access-l8kxb") pod "9ce6139b-7b9f-449e-a642-ca71b4740873" (UID: "9ce6139b-7b9f-449e-a642-ca71b4740873"). InnerVolumeSpecName "kube-api-access-l8kxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.762866 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-config" (OuterVolumeSpecName: "config") pod "9ce6139b-7b9f-449e-a642-ca71b4740873" (UID: "9ce6139b-7b9f-449e-a642-ca71b4740873"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.766051 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9ce6139b-7b9f-449e-a642-ca71b4740873" (UID: "9ce6139b-7b9f-449e-a642-ca71b4740873"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.767862 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9ce6139b-7b9f-449e-a642-ca71b4740873" (UID: "9ce6139b-7b9f-449e-a642-ca71b4740873"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.774140 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9ce6139b-7b9f-449e-a642-ca71b4740873" (UID: "9ce6139b-7b9f-449e-a642-ca71b4740873"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.820241 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.820288 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.820302 4739 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.820314 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8kxb\" (UniqueName: \"kubernetes.io/projected/9ce6139b-7b9f-449e-a642-ca71b4740873-kube-api-access-l8kxb\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.820326 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ce6139b-7b9f-449e-a642-ca71b4740873-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.953433 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-vzh4g"] Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.962919 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b946d459c-vzh4g"] Dec 01 15:53:42 crc kubenswrapper[4739]: I1201 15:53:42.964969 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5d98bc475f-59p2r" Dec 01 15:53:43 crc kubenswrapper[4739]: I1201 15:53:43.023911 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7d77968bfd-m6mqb"] Dec 01 15:53:43 crc kubenswrapper[4739]: I1201 15:53:43.024304 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7d77968bfd-m6mqb" podUID="fa675bf0-e417-4847-ad0b-4f364208d007" containerName="neutron-api" containerID="cri-o://ba9b34bec7c662fef25c301e6be902f932f1082246a59b4dd662c2fc778b9706" gracePeriod=30 Dec 01 15:53:43 crc kubenswrapper[4739]: I1201 15:53:43.024719 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7d77968bfd-m6mqb" podUID="fa675bf0-e417-4847-ad0b-4f364208d007" containerName="neutron-httpd" containerID="cri-o://bfdf2eb822c5bb2516c0f817fac22d3ae840d3cdcfd06762d0f4823b9250636a" gracePeriod=30 Dec 01 15:53:43 crc kubenswrapper[4739]: I1201 15:53:43.633103 4739 generic.go:334] "Generic (PLEG): container finished" podID="76ccf696-2126-4fa1-b6ec-1a35c1251e90" containerID="12e3e6f9a0de70e3cb685e1c27e2d3b556e71dd38a6c45ad5b4d57ae9e923208" exitCode=0 Dec 01 15:53:43 crc kubenswrapper[4739]: I1201 15:53:43.633181 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"76ccf696-2126-4fa1-b6ec-1a35c1251e90","Type":"ContainerDied","Data":"12e3e6f9a0de70e3cb685e1c27e2d3b556e71dd38a6c45ad5b4d57ae9e923208"} Dec 01 15:53:43 crc kubenswrapper[4739]: I1201 15:53:43.635258 4739 generic.go:334] "Generic (PLEG): container finished" podID="fa675bf0-e417-4847-ad0b-4f364208d007" containerID="bfdf2eb822c5bb2516c0f817fac22d3ae840d3cdcfd06762d0f4823b9250636a" exitCode=0 Dec 01 15:53:43 crc kubenswrapper[4739]: I1201 15:53:43.635323 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d77968bfd-m6mqb" event={"ID":"fa675bf0-e417-4847-ad0b-4f364208d007","Type":"ContainerDied","Data":"bfdf2eb822c5bb2516c0f817fac22d3ae840d3cdcfd06762d0f4823b9250636a"} Dec 01 15:53:44 crc kubenswrapper[4739]: I1201 15:53:44.493590 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ce6139b-7b9f-449e-a642-ca71b4740873" path="/var/lib/kubelet/pods/9ce6139b-7b9f-449e-a642-ca71b4740873/volumes" Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.419320 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.469155 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-54db66b7f6-6plgv" Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.662566 4739 generic.go:334] "Generic (PLEG): container finished" podID="76ccf696-2126-4fa1-b6ec-1a35c1251e90" containerID="0963a9c722c8dcb52061f80b83ed810edc4e8229ab029f83e7bec46983a3ed5f" exitCode=0 Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.663352 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"76ccf696-2126-4fa1-b6ec-1a35c1251e90","Type":"ContainerDied","Data":"0963a9c722c8dcb52061f80b83ed810edc4e8229ab029f83e7bec46983a3ed5f"} Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.730272 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.882260 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-scripts\") pod \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.882313 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqrn5\" (UniqueName: \"kubernetes.io/projected/76ccf696-2126-4fa1-b6ec-1a35c1251e90-kube-api-access-cqrn5\") pod \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.882348 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-combined-ca-bundle\") pod \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.882384 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-config-data-custom\") pod \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.882401 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-config-data\") pod \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.882551 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76ccf696-2126-4fa1-b6ec-1a35c1251e90-etc-machine-id\") pod \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\" (UID: \"76ccf696-2126-4fa1-b6ec-1a35c1251e90\") " Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.882895 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/76ccf696-2126-4fa1-b6ec-1a35c1251e90-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "76ccf696-2126-4fa1-b6ec-1a35c1251e90" (UID: "76ccf696-2126-4fa1-b6ec-1a35c1251e90"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.888526 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76ccf696-2126-4fa1-b6ec-1a35c1251e90-kube-api-access-cqrn5" (OuterVolumeSpecName: "kube-api-access-cqrn5") pod "76ccf696-2126-4fa1-b6ec-1a35c1251e90" (UID: "76ccf696-2126-4fa1-b6ec-1a35c1251e90"). InnerVolumeSpecName "kube-api-access-cqrn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.888964 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-scripts" (OuterVolumeSpecName: "scripts") pod "76ccf696-2126-4fa1-b6ec-1a35c1251e90" (UID: "76ccf696-2126-4fa1-b6ec-1a35c1251e90"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.902086 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "76ccf696-2126-4fa1-b6ec-1a35c1251e90" (UID: "76ccf696-2126-4fa1-b6ec-1a35c1251e90"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.930189 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76ccf696-2126-4fa1-b6ec-1a35c1251e90" (UID: "76ccf696-2126-4fa1-b6ec-1a35c1251e90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.981387 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-config-data" (OuterVolumeSpecName: "config-data") pod "76ccf696-2126-4fa1-b6ec-1a35c1251e90" (UID: "76ccf696-2126-4fa1-b6ec-1a35c1251e90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.984791 4739 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.984839 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.984858 4739 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76ccf696-2126-4fa1-b6ec-1a35c1251e90-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.984875 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.984892 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqrn5\" (UniqueName: \"kubernetes.io/projected/76ccf696-2126-4fa1-b6ec-1a35c1251e90-kube-api-access-cqrn5\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:45 crc kubenswrapper[4739]: I1201 15:53:45.984911 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ccf696-2126-4fa1-b6ec-1a35c1251e90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.675008 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"76ccf696-2126-4fa1-b6ec-1a35c1251e90","Type":"ContainerDied","Data":"7edab46af5fae2b786a1e90baf91a7bf63a22ec5ac252d8cc9859671304cb8cf"} Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.675157 4739 scope.go:117] "RemoveContainer" containerID="12e3e6f9a0de70e3cb685e1c27e2d3b556e71dd38a6c45ad5b4d57ae9e923208" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.675247 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.701117 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.718083 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.733888 4739 scope.go:117] "RemoveContainer" containerID="0963a9c722c8dcb52061f80b83ed810edc4e8229ab029f83e7bec46983a3ed5f" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.734310 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 15:53:46 crc kubenswrapper[4739]: E1201 15:53:46.734760 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ce6139b-7b9f-449e-a642-ca71b4740873" containerName="dnsmasq-dns" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.734784 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ce6139b-7b9f-449e-a642-ca71b4740873" containerName="dnsmasq-dns" Dec 01 15:53:46 crc kubenswrapper[4739]: E1201 15:53:46.734806 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ccf696-2126-4fa1-b6ec-1a35c1251e90" containerName="cinder-scheduler" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.734813 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ccf696-2126-4fa1-b6ec-1a35c1251e90" containerName="cinder-scheduler" Dec 01 15:53:46 crc kubenswrapper[4739]: E1201 15:53:46.734822 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ccf696-2126-4fa1-b6ec-1a35c1251e90" containerName="probe" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.734829 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ccf696-2126-4fa1-b6ec-1a35c1251e90" containerName="probe" Dec 01 15:53:46 crc kubenswrapper[4739]: E1201 15:53:46.734860 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ce6139b-7b9f-449e-a642-ca71b4740873" containerName="init" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.734868 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ce6139b-7b9f-449e-a642-ca71b4740873" containerName="init" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.735057 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="76ccf696-2126-4fa1-b6ec-1a35c1251e90" containerName="cinder-scheduler" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.735084 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="76ccf696-2126-4fa1-b6ec-1a35c1251e90" containerName="probe" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.735231 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ce6139b-7b9f-449e-a642-ca71b4740873" containerName="dnsmasq-dns" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.736289 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.739402 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.748906 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.907791 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.907867 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98v87\" (UniqueName: \"kubernetes.io/projected/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-kube-api-access-98v87\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.907915 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.908001 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.908040 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-config-data\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:46 crc kubenswrapper[4739]: I1201 15:53:46.908061 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-scripts\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.009962 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.010057 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.010074 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-config-data\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.010092 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-scripts\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.010152 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.010170 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98v87\" (UniqueName: \"kubernetes.io/projected/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-kube-api-access-98v87\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.011437 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.014300 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-scripts\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.015319 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.015371 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-config-data\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.018985 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.031682 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98v87\" (UniqueName: \"kubernetes.io/projected/4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2-kube-api-access-98v87\") pod \"cinder-scheduler-0\" (UID: \"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2\") " pod="openstack/cinder-scheduler-0" Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.071981 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.542109 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.690256 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2","Type":"ContainerStarted","Data":"8a07c3567dce3028295bbbe8ea08560baf326c93e6292731209bf684393036e1"} Dec 01 15:53:47 crc kubenswrapper[4739]: I1201 15:53:47.991052 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:48 crc kubenswrapper[4739]: I1201 15:53:48.169374 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-856bfcfff6-5wtcc" Dec 01 15:53:48 crc kubenswrapper[4739]: I1201 15:53:48.255682 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-799f5fcdc4-bwhj9"] Dec 01 15:53:48 crc kubenswrapper[4739]: I1201 15:53:48.255947 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-799f5fcdc4-bwhj9" podUID="3703361d-b3f1-4139-85da-cf21e825b7f2" containerName="barbican-api-log" containerID="cri-o://de3766274b2e61080a4ae14648f3d44fc5582e6ba32fbcfb4278277b14bcdadc" gracePeriod=30 Dec 01 15:53:48 crc kubenswrapper[4739]: I1201 15:53:48.256282 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-799f5fcdc4-bwhj9" podUID="3703361d-b3f1-4139-85da-cf21e825b7f2" containerName="barbican-api" containerID="cri-o://cb69ae7e451089101b165f5812a20a66fbcff3a736c206834d1860d6bf3bd0f8" gracePeriod=30 Dec 01 15:53:48 crc kubenswrapper[4739]: I1201 15:53:48.500065 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76ccf696-2126-4fa1-b6ec-1a35c1251e90" path="/var/lib/kubelet/pods/76ccf696-2126-4fa1-b6ec-1a35c1251e90/volumes" Dec 01 15:53:48 crc kubenswrapper[4739]: I1201 15:53:48.705776 4739 generic.go:334] "Generic (PLEG): container finished" podID="3703361d-b3f1-4139-85da-cf21e825b7f2" containerID="de3766274b2e61080a4ae14648f3d44fc5582e6ba32fbcfb4278277b14bcdadc" exitCode=143 Dec 01 15:53:48 crc kubenswrapper[4739]: I1201 15:53:48.705817 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-799f5fcdc4-bwhj9" event={"ID":"3703361d-b3f1-4139-85da-cf21e825b7f2","Type":"ContainerDied","Data":"de3766274b2e61080a4ae14648f3d44fc5582e6ba32fbcfb4278277b14bcdadc"} Dec 01 15:53:48 crc kubenswrapper[4739]: I1201 15:53:48.708502 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2","Type":"ContainerStarted","Data":"b3f023714c8c5367af00bcfe628a260ee37e04d259007cd315bfaa930e8e93f5"} Dec 01 15:53:49 crc kubenswrapper[4739]: I1201 15:53:49.743513 4739 generic.go:334] "Generic (PLEG): container finished" podID="fa675bf0-e417-4847-ad0b-4f364208d007" containerID="ba9b34bec7c662fef25c301e6be902f932f1082246a59b4dd662c2fc778b9706" exitCode=0 Dec 01 15:53:49 crc kubenswrapper[4739]: I1201 15:53:49.743893 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d77968bfd-m6mqb" event={"ID":"fa675bf0-e417-4847-ad0b-4f364208d007","Type":"ContainerDied","Data":"ba9b34bec7c662fef25c301e6be902f932f1082246a59b4dd662c2fc778b9706"} Dec 01 15:53:49 crc kubenswrapper[4739]: I1201 15:53:49.745593 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2","Type":"ContainerStarted","Data":"8b0e2f14cae31f142fa4bbc0067ea3dc4b42ed9f30b051362308a9f33915b879"} Dec 01 15:53:49 crc kubenswrapper[4739]: I1201 15:53:49.775912 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.7758933519999998 podStartE2EDuration="3.775893352s" podCreationTimestamp="2025-12-01 15:53:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:53:49.771202097 +0000 UTC m=+1131.596948201" watchObservedRunningTime="2025-12-01 15:53:49.775893352 +0000 UTC m=+1131.601639456" Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.116134 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.268550 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-combined-ca-bundle\") pod \"fa675bf0-e417-4847-ad0b-4f364208d007\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.268883 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-ovndb-tls-certs\") pod \"fa675bf0-e417-4847-ad0b-4f364208d007\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.268930 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-httpd-config\") pod \"fa675bf0-e417-4847-ad0b-4f364208d007\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.268949 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-config\") pod \"fa675bf0-e417-4847-ad0b-4f364208d007\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.269135 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdv2c\" (UniqueName: \"kubernetes.io/projected/fa675bf0-e417-4847-ad0b-4f364208d007-kube-api-access-qdv2c\") pod \"fa675bf0-e417-4847-ad0b-4f364208d007\" (UID: \"fa675bf0-e417-4847-ad0b-4f364208d007\") " Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.274031 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa675bf0-e417-4847-ad0b-4f364208d007-kube-api-access-qdv2c" (OuterVolumeSpecName: "kube-api-access-qdv2c") pod "fa675bf0-e417-4847-ad0b-4f364208d007" (UID: "fa675bf0-e417-4847-ad0b-4f364208d007"). InnerVolumeSpecName "kube-api-access-qdv2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.282796 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "fa675bf0-e417-4847-ad0b-4f364208d007" (UID: "fa675bf0-e417-4847-ad0b-4f364208d007"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.315604 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-config" (OuterVolumeSpecName: "config") pod "fa675bf0-e417-4847-ad0b-4f364208d007" (UID: "fa675bf0-e417-4847-ad0b-4f364208d007"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.324561 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fa675bf0-e417-4847-ad0b-4f364208d007" (UID: "fa675bf0-e417-4847-ad0b-4f364208d007"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.337479 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "fa675bf0-e417-4847-ad0b-4f364208d007" (UID: "fa675bf0-e417-4847-ad0b-4f364208d007"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.370872 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.371963 4739 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.372165 4739 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.372295 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/fa675bf0-e417-4847-ad0b-4f364208d007-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.373656 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdv2c\" (UniqueName: \"kubernetes.io/projected/fa675bf0-e417-4847-ad0b-4f364208d007-kube-api-access-qdv2c\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.756840 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d77968bfd-m6mqb" Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.756891 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d77968bfd-m6mqb" event={"ID":"fa675bf0-e417-4847-ad0b-4f364208d007","Type":"ContainerDied","Data":"8e4ab5bcb7ea05843314b1d96d8f250bd50647a4576ceedccfd7e6e146afb2b0"} Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.756929 4739 scope.go:117] "RemoveContainer" containerID="bfdf2eb822c5bb2516c0f817fac22d3ae840d3cdcfd06762d0f4823b9250636a" Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.792656 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7d77968bfd-m6mqb"] Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.802489 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7d77968bfd-m6mqb"] Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.806671 4739 scope.go:117] "RemoveContainer" containerID="ba9b34bec7c662fef25c301e6be902f932f1082246a59b4dd662c2fc778b9706" Dec 01 15:53:50 crc kubenswrapper[4739]: I1201 15:53:50.909361 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 01 15:53:51 crc kubenswrapper[4739]: I1201 15:53:51.418669 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-799f5fcdc4-bwhj9" podUID="3703361d-b3f1-4139-85da-cf21e825b7f2" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.147:9311/healthcheck\": read tcp 10.217.0.2:58406->10.217.0.147:9311: read: connection reset by peer" Dec 01 15:53:51 crc kubenswrapper[4739]: I1201 15:53:51.418769 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-799f5fcdc4-bwhj9" podUID="3703361d-b3f1-4139-85da-cf21e825b7f2" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.147:9311/healthcheck\": read tcp 10.217.0.2:58408->10.217.0.147:9311: read: connection reset by peer" Dec 01 15:53:51 crc kubenswrapper[4739]: I1201 15:53:51.769325 4739 generic.go:334] "Generic (PLEG): container finished" podID="3703361d-b3f1-4139-85da-cf21e825b7f2" containerID="cb69ae7e451089101b165f5812a20a66fbcff3a736c206834d1860d6bf3bd0f8" exitCode=0 Dec 01 15:53:51 crc kubenswrapper[4739]: I1201 15:53:51.769482 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-799f5fcdc4-bwhj9" event={"ID":"3703361d-b3f1-4139-85da-cf21e825b7f2","Type":"ContainerDied","Data":"cb69ae7e451089101b165f5812a20a66fbcff3a736c206834d1860d6bf3bd0f8"} Dec 01 15:53:51 crc kubenswrapper[4739]: I1201 15:53:51.859147 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.014224 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-config-data\") pod \"3703361d-b3f1-4139-85da-cf21e825b7f2\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.014317 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgpvg\" (UniqueName: \"kubernetes.io/projected/3703361d-b3f1-4139-85da-cf21e825b7f2-kube-api-access-pgpvg\") pod \"3703361d-b3f1-4139-85da-cf21e825b7f2\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.014408 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-config-data-custom\") pod \"3703361d-b3f1-4139-85da-cf21e825b7f2\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.014645 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3703361d-b3f1-4139-85da-cf21e825b7f2-logs\") pod \"3703361d-b3f1-4139-85da-cf21e825b7f2\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.014729 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-combined-ca-bundle\") pod \"3703361d-b3f1-4139-85da-cf21e825b7f2\" (UID: \"3703361d-b3f1-4139-85da-cf21e825b7f2\") " Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.015661 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3703361d-b3f1-4139-85da-cf21e825b7f2-logs" (OuterVolumeSpecName: "logs") pod "3703361d-b3f1-4139-85da-cf21e825b7f2" (UID: "3703361d-b3f1-4139-85da-cf21e825b7f2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.016479 4739 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3703361d-b3f1-4139-85da-cf21e825b7f2-logs\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.023235 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3703361d-b3f1-4139-85da-cf21e825b7f2-kube-api-access-pgpvg" (OuterVolumeSpecName: "kube-api-access-pgpvg") pod "3703361d-b3f1-4139-85da-cf21e825b7f2" (UID: "3703361d-b3f1-4139-85da-cf21e825b7f2"). InnerVolumeSpecName "kube-api-access-pgpvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.028869 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3703361d-b3f1-4139-85da-cf21e825b7f2" (UID: "3703361d-b3f1-4139-85da-cf21e825b7f2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.072669 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.075537 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3703361d-b3f1-4139-85da-cf21e825b7f2" (UID: "3703361d-b3f1-4139-85da-cf21e825b7f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.114272 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-config-data" (OuterVolumeSpecName: "config-data") pod "3703361d-b3f1-4139-85da-cf21e825b7f2" (UID: "3703361d-b3f1-4139-85da-cf21e825b7f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.118440 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.118471 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgpvg\" (UniqueName: \"kubernetes.io/projected/3703361d-b3f1-4139-85da-cf21e825b7f2-kube-api-access-pgpvg\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.118487 4739 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.118500 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3703361d-b3f1-4139-85da-cf21e825b7f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.490251 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa675bf0-e417-4847-ad0b-4f364208d007" path="/var/lib/kubelet/pods/fa675bf0-e417-4847-ad0b-4f364208d007/volumes" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.787172 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-799f5fcdc4-bwhj9" event={"ID":"3703361d-b3f1-4139-85da-cf21e825b7f2","Type":"ContainerDied","Data":"4c0bce3f28acd172342cc7f86639441c813e5e9ebe707209aeee590004b8c579"} Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.787614 4739 scope.go:117] "RemoveContainer" containerID="cb69ae7e451089101b165f5812a20a66fbcff3a736c206834d1860d6bf3bd0f8" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.787573 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-799f5fcdc4-bwhj9" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.818914 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-799f5fcdc4-bwhj9"] Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.831035 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-799f5fcdc4-bwhj9"] Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.834756 4739 scope.go:117] "RemoveContainer" containerID="de3766274b2e61080a4ae14648f3d44fc5582e6ba32fbcfb4278277b14bcdadc" Dec 01 15:53:52 crc kubenswrapper[4739]: I1201 15:53:52.894493 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5f489c7bf6-cm54s" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.493217 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3703361d-b3f1-4139-85da-cf21e825b7f2" path="/var/lib/kubelet/pods/3703361d-b3f1-4139-85da-cf21e825b7f2/volumes" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.589110 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 15:53:54 crc kubenswrapper[4739]: E1201 15:53:54.589554 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3703361d-b3f1-4139-85da-cf21e825b7f2" containerName="barbican-api-log" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.589570 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="3703361d-b3f1-4139-85da-cf21e825b7f2" containerName="barbican-api-log" Dec 01 15:53:54 crc kubenswrapper[4739]: E1201 15:53:54.589584 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3703361d-b3f1-4139-85da-cf21e825b7f2" containerName="barbican-api" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.589590 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="3703361d-b3f1-4139-85da-cf21e825b7f2" containerName="barbican-api" Dec 01 15:53:54 crc kubenswrapper[4739]: E1201 15:53:54.589609 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa675bf0-e417-4847-ad0b-4f364208d007" containerName="neutron-httpd" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.589616 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa675bf0-e417-4847-ad0b-4f364208d007" containerName="neutron-httpd" Dec 01 15:53:54 crc kubenswrapper[4739]: E1201 15:53:54.589630 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa675bf0-e417-4847-ad0b-4f364208d007" containerName="neutron-api" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.589635 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa675bf0-e417-4847-ad0b-4f364208d007" containerName="neutron-api" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.589866 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa675bf0-e417-4847-ad0b-4f364208d007" containerName="neutron-api" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.589881 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="3703361d-b3f1-4139-85da-cf21e825b7f2" containerName="barbican-api-log" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.589894 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="3703361d-b3f1-4139-85da-cf21e825b7f2" containerName="barbican-api" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.589905 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa675bf0-e417-4847-ad0b-4f364208d007" containerName="neutron-httpd" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.590508 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.598212 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.599815 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-gwrcd" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.600053 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.601339 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.768493 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a2651dc-3474-4da9-8cc8-187c8a392203-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " pod="openstack/openstackclient" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.768560 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a2651dc-3474-4da9-8cc8-187c8a392203-openstack-config-secret\") pod \"openstackclient\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " pod="openstack/openstackclient" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.768609 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj767\" (UniqueName: \"kubernetes.io/projected/6a2651dc-3474-4da9-8cc8-187c8a392203-kube-api-access-jj767\") pod \"openstackclient\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " pod="openstack/openstackclient" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.768692 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a2651dc-3474-4da9-8cc8-187c8a392203-openstack-config\") pod \"openstackclient\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " pod="openstack/openstackclient" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.870332 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a2651dc-3474-4da9-8cc8-187c8a392203-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " pod="openstack/openstackclient" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.870454 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a2651dc-3474-4da9-8cc8-187c8a392203-openstack-config-secret\") pod \"openstackclient\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " pod="openstack/openstackclient" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.870529 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj767\" (UniqueName: \"kubernetes.io/projected/6a2651dc-3474-4da9-8cc8-187c8a392203-kube-api-access-jj767\") pod \"openstackclient\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " pod="openstack/openstackclient" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.870623 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a2651dc-3474-4da9-8cc8-187c8a392203-openstack-config\") pod \"openstackclient\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " pod="openstack/openstackclient" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.871475 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a2651dc-3474-4da9-8cc8-187c8a392203-openstack-config\") pod \"openstackclient\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " pod="openstack/openstackclient" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.877882 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a2651dc-3474-4da9-8cc8-187c8a392203-openstack-config-secret\") pod \"openstackclient\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " pod="openstack/openstackclient" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.889243 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj767\" (UniqueName: \"kubernetes.io/projected/6a2651dc-3474-4da9-8cc8-187c8a392203-kube-api-access-jj767\") pod \"openstackclient\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " pod="openstack/openstackclient" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.897103 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a2651dc-3474-4da9-8cc8-187c8a392203-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " pod="openstack/openstackclient" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.954527 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 15:53:54 crc kubenswrapper[4739]: I1201 15:53:54.999604 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.008296 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.084067 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.085483 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.101016 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 15:53:55 crc kubenswrapper[4739]: E1201 15:53:55.152947 4739 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 01 15:53:55 crc kubenswrapper[4739]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_6a2651dc-3474-4da9-8cc8-187c8a392203_0(bddd75d1c425e9886167201b2961364b8f10b1a2865e9d28d5b9f184fecdb5bd): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"bddd75d1c425e9886167201b2961364b8f10b1a2865e9d28d5b9f184fecdb5bd" Netns:"/var/run/netns/f1afe73a-a4f8-4a16-b796-dd1be8e972e0" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=bddd75d1c425e9886167201b2961364b8f10b1a2865e9d28d5b9f184fecdb5bd;K8S_POD_UID=6a2651dc-3474-4da9-8cc8-187c8a392203" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/6a2651dc-3474-4da9-8cc8-187c8a392203]: expected pod UID "6a2651dc-3474-4da9-8cc8-187c8a392203" but got "ef1ce8cc-853d-415c-85be-39fc10bec74d" from Kube API Dec 01 15:53:55 crc kubenswrapper[4739]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 01 15:53:55 crc kubenswrapper[4739]: > Dec 01 15:53:55 crc kubenswrapper[4739]: E1201 15:53:55.153015 4739 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 01 15:53:55 crc kubenswrapper[4739]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_6a2651dc-3474-4da9-8cc8-187c8a392203_0(bddd75d1c425e9886167201b2961364b8f10b1a2865e9d28d5b9f184fecdb5bd): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"bddd75d1c425e9886167201b2961364b8f10b1a2865e9d28d5b9f184fecdb5bd" Netns:"/var/run/netns/f1afe73a-a4f8-4a16-b796-dd1be8e972e0" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=bddd75d1c425e9886167201b2961364b8f10b1a2865e9d28d5b9f184fecdb5bd;K8S_POD_UID=6a2651dc-3474-4da9-8cc8-187c8a392203" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/6a2651dc-3474-4da9-8cc8-187c8a392203]: expected pod UID "6a2651dc-3474-4da9-8cc8-187c8a392203" but got "ef1ce8cc-853d-415c-85be-39fc10bec74d" from Kube API Dec 01 15:53:55 crc kubenswrapper[4739]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 01 15:53:55 crc kubenswrapper[4739]: > pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.177207 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ef1ce8cc-853d-415c-85be-39fc10bec74d-openstack-config\") pod \"openstackclient\" (UID: \"ef1ce8cc-853d-415c-85be-39fc10bec74d\") " pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.177311 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ef1ce8cc-853d-415c-85be-39fc10bec74d-openstack-config-secret\") pod \"openstackclient\" (UID: \"ef1ce8cc-853d-415c-85be-39fc10bec74d\") " pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.177340 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2thrx\" (UniqueName: \"kubernetes.io/projected/ef1ce8cc-853d-415c-85be-39fc10bec74d-kube-api-access-2thrx\") pod \"openstackclient\" (UID: \"ef1ce8cc-853d-415c-85be-39fc10bec74d\") " pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.177526 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef1ce8cc-853d-415c-85be-39fc10bec74d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ef1ce8cc-853d-415c-85be-39fc10bec74d\") " pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.279049 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef1ce8cc-853d-415c-85be-39fc10bec74d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ef1ce8cc-853d-415c-85be-39fc10bec74d\") " pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.279094 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ef1ce8cc-853d-415c-85be-39fc10bec74d-openstack-config\") pod \"openstackclient\" (UID: \"ef1ce8cc-853d-415c-85be-39fc10bec74d\") " pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.279175 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ef1ce8cc-853d-415c-85be-39fc10bec74d-openstack-config-secret\") pod \"openstackclient\" (UID: \"ef1ce8cc-853d-415c-85be-39fc10bec74d\") " pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.280056 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ef1ce8cc-853d-415c-85be-39fc10bec74d-openstack-config\") pod \"openstackclient\" (UID: \"ef1ce8cc-853d-415c-85be-39fc10bec74d\") " pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.280103 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2thrx\" (UniqueName: \"kubernetes.io/projected/ef1ce8cc-853d-415c-85be-39fc10bec74d-kube-api-access-2thrx\") pod \"openstackclient\" (UID: \"ef1ce8cc-853d-415c-85be-39fc10bec74d\") " pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.284443 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ef1ce8cc-853d-415c-85be-39fc10bec74d-openstack-config-secret\") pod \"openstackclient\" (UID: \"ef1ce8cc-853d-415c-85be-39fc10bec74d\") " pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.285651 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef1ce8cc-853d-415c-85be-39fc10bec74d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ef1ce8cc-853d-415c-85be-39fc10bec74d\") " pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.294762 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2thrx\" (UniqueName: \"kubernetes.io/projected/ef1ce8cc-853d-415c-85be-39fc10bec74d-kube-api-access-2thrx\") pod \"openstackclient\" (UID: \"ef1ce8cc-853d-415c-85be-39fc10bec74d\") " pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.410866 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.822208 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.825290 4739 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="6a2651dc-3474-4da9-8cc8-187c8a392203" podUID="ef1ce8cc-853d-415c-85be-39fc10bec74d" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.838087 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.882864 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 15:53:55 crc kubenswrapper[4739]: W1201 15:53:55.888454 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef1ce8cc_853d_415c_85be_39fc10bec74d.slice/crio-760f8d77b2c2defd7be296b14c858030e90e12259e3e422ed61e8d6cef83af9f WatchSource:0}: Error finding container 760f8d77b2c2defd7be296b14c858030e90e12259e3e422ed61e8d6cef83af9f: Status 404 returned error can't find the container with id 760f8d77b2c2defd7be296b14c858030e90e12259e3e422ed61e8d6cef83af9f Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.990692 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a2651dc-3474-4da9-8cc8-187c8a392203-combined-ca-bundle\") pod \"6a2651dc-3474-4da9-8cc8-187c8a392203\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.990869 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a2651dc-3474-4da9-8cc8-187c8a392203-openstack-config\") pod \"6a2651dc-3474-4da9-8cc8-187c8a392203\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.990922 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a2651dc-3474-4da9-8cc8-187c8a392203-openstack-config-secret\") pod \"6a2651dc-3474-4da9-8cc8-187c8a392203\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.991156 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jj767\" (UniqueName: \"kubernetes.io/projected/6a2651dc-3474-4da9-8cc8-187c8a392203-kube-api-access-jj767\") pod \"6a2651dc-3474-4da9-8cc8-187c8a392203\" (UID: \"6a2651dc-3474-4da9-8cc8-187c8a392203\") " Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.994615 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a2651dc-3474-4da9-8cc8-187c8a392203-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "6a2651dc-3474-4da9-8cc8-187c8a392203" (UID: "6a2651dc-3474-4da9-8cc8-187c8a392203"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.998526 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a2651dc-3474-4da9-8cc8-187c8a392203-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a2651dc-3474-4da9-8cc8-187c8a392203" (UID: "6a2651dc-3474-4da9-8cc8-187c8a392203"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:55 crc kubenswrapper[4739]: I1201 15:53:55.998566 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a2651dc-3474-4da9-8cc8-187c8a392203-kube-api-access-jj767" (OuterVolumeSpecName: "kube-api-access-jj767") pod "6a2651dc-3474-4da9-8cc8-187c8a392203" (UID: "6a2651dc-3474-4da9-8cc8-187c8a392203"). InnerVolumeSpecName "kube-api-access-jj767". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:56 crc kubenswrapper[4739]: I1201 15:53:56.001577 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a2651dc-3474-4da9-8cc8-187c8a392203-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "6a2651dc-3474-4da9-8cc8-187c8a392203" (UID: "6a2651dc-3474-4da9-8cc8-187c8a392203"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:53:56 crc kubenswrapper[4739]: I1201 15:53:56.093084 4739 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a2651dc-3474-4da9-8cc8-187c8a392203-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:56 crc kubenswrapper[4739]: I1201 15:53:56.093135 4739 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a2651dc-3474-4da9-8cc8-187c8a392203-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:56 crc kubenswrapper[4739]: I1201 15:53:56.093150 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj767\" (UniqueName: \"kubernetes.io/projected/6a2651dc-3474-4da9-8cc8-187c8a392203-kube-api-access-jj767\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:56 crc kubenswrapper[4739]: I1201 15:53:56.093161 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a2651dc-3474-4da9-8cc8-187c8a392203-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:56 crc kubenswrapper[4739]: I1201 15:53:56.487795 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a2651dc-3474-4da9-8cc8-187c8a392203" path="/var/lib/kubelet/pods/6a2651dc-3474-4da9-8cc8-187c8a392203/volumes" Dec 01 15:53:56 crc kubenswrapper[4739]: I1201 15:53:56.829998 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 15:53:56 crc kubenswrapper[4739]: I1201 15:53:56.830368 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ef1ce8cc-853d-415c-85be-39fc10bec74d","Type":"ContainerStarted","Data":"760f8d77b2c2defd7be296b14c858030e90e12259e3e422ed61e8d6cef83af9f"} Dec 01 15:53:56 crc kubenswrapper[4739]: I1201 15:53:56.836319 4739 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="6a2651dc-3474-4da9-8cc8-187c8a392203" podUID="ef1ce8cc-853d-415c-85be-39fc10bec74d" Dec 01 15:53:57 crc kubenswrapper[4739]: I1201 15:53:57.363754 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 15:54:00 crc kubenswrapper[4739]: I1201 15:54:00.026569 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 15:54:00 crc kubenswrapper[4739]: I1201 15:54:00.188258 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:54:00 crc kubenswrapper[4739]: I1201 15:54:00.862166 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="ceilometer-central-agent" containerID="cri-o://caf209b4d2ee1eb81d3f05a38cdc7cf0fa764be103faeb2ea353a40594224d76" gracePeriod=30 Dec 01 15:54:00 crc kubenswrapper[4739]: I1201 15:54:00.862241 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="proxy-httpd" containerID="cri-o://3d096b0f728cd7c50a87d1e77a88cd58239e705a29aaa6fa8ee016ed1cfdb9a9" gracePeriod=30 Dec 01 15:54:00 crc kubenswrapper[4739]: I1201 15:54:00.862257 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="ceilometer-notification-agent" containerID="cri-o://d34f390505dfaf24ddeea1b458f02ef3eb977f5d31823a8f72cd3ef73d02b41f" gracePeriod=30 Dec 01 15:54:00 crc kubenswrapper[4739]: I1201 15:54:00.862241 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="sg-core" containerID="cri-o://d3bb11e2c8f408e8499881b7bb9857a7ad272ecb71d08fcc893c02aeb5afb5b4" gracePeriod=30 Dec 01 15:54:01 crc kubenswrapper[4739]: I1201 15:54:01.874615 4739 generic.go:334] "Generic (PLEG): container finished" podID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerID="3d096b0f728cd7c50a87d1e77a88cd58239e705a29aaa6fa8ee016ed1cfdb9a9" exitCode=0 Dec 01 15:54:01 crc kubenswrapper[4739]: I1201 15:54:01.874833 4739 generic.go:334] "Generic (PLEG): container finished" podID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerID="d3bb11e2c8f408e8499881b7bb9857a7ad272ecb71d08fcc893c02aeb5afb5b4" exitCode=2 Dec 01 15:54:01 crc kubenswrapper[4739]: I1201 15:54:01.874844 4739 generic.go:334] "Generic (PLEG): container finished" podID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerID="caf209b4d2ee1eb81d3f05a38cdc7cf0fa764be103faeb2ea353a40594224d76" exitCode=0 Dec 01 15:54:01 crc kubenswrapper[4739]: I1201 15:54:01.874754 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d79c5700-e1a6-416b-bba1-5c211439c9d9","Type":"ContainerDied","Data":"3d096b0f728cd7c50a87d1e77a88cd58239e705a29aaa6fa8ee016ed1cfdb9a9"} Dec 01 15:54:01 crc kubenswrapper[4739]: I1201 15:54:01.874869 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d79c5700-e1a6-416b-bba1-5c211439c9d9","Type":"ContainerDied","Data":"d3bb11e2c8f408e8499881b7bb9857a7ad272ecb71d08fcc893c02aeb5afb5b4"} Dec 01 15:54:01 crc kubenswrapper[4739]: I1201 15:54:01.874880 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d79c5700-e1a6-416b-bba1-5c211439c9d9","Type":"ContainerDied","Data":"caf209b4d2ee1eb81d3f05a38cdc7cf0fa764be103faeb2ea353a40594224d76"} Dec 01 15:54:04 crc kubenswrapper[4739]: I1201 15:54:04.912990 4739 generic.go:334] "Generic (PLEG): container finished" podID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerID="d34f390505dfaf24ddeea1b458f02ef3eb977f5d31823a8f72cd3ef73d02b41f" exitCode=0 Dec 01 15:54:04 crc kubenswrapper[4739]: I1201 15:54:04.913515 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d79c5700-e1a6-416b-bba1-5c211439c9d9","Type":"ContainerDied","Data":"d34f390505dfaf24ddeea1b458f02ef3eb977f5d31823a8f72cd3ef73d02b41f"} Dec 01 15:54:04 crc kubenswrapper[4739]: I1201 15:54:04.928298 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ef1ce8cc-853d-415c-85be-39fc10bec74d","Type":"ContainerStarted","Data":"da21b29763e40590164b4467946a619f0a30fa177f56ad85bb22941b2d507086"} Dec 01 15:54:04 crc kubenswrapper[4739]: I1201 15:54:04.945541 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.3065168090000001 podStartE2EDuration="9.94552414s" podCreationTimestamp="2025-12-01 15:53:55 +0000 UTC" firstStartedPulling="2025-12-01 15:53:55.89124937 +0000 UTC m=+1137.716995484" lastFinishedPulling="2025-12-01 15:54:04.530256721 +0000 UTC m=+1146.356002815" observedRunningTime="2025-12-01 15:54:04.943823738 +0000 UTC m=+1146.769569832" watchObservedRunningTime="2025-12-01 15:54:04.94552414 +0000 UTC m=+1146.771270234" Dec 01 15:54:04 crc kubenswrapper[4739]: I1201 15:54:04.991526 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.172585 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-sg-core-conf-yaml\") pod \"d79c5700-e1a6-416b-bba1-5c211439c9d9\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.172652 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-scripts\") pod \"d79c5700-e1a6-416b-bba1-5c211439c9d9\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.172777 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r669x\" (UniqueName: \"kubernetes.io/projected/d79c5700-e1a6-416b-bba1-5c211439c9d9-kube-api-access-r669x\") pod \"d79c5700-e1a6-416b-bba1-5c211439c9d9\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.172813 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d79c5700-e1a6-416b-bba1-5c211439c9d9-log-httpd\") pod \"d79c5700-e1a6-416b-bba1-5c211439c9d9\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.172845 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-combined-ca-bundle\") pod \"d79c5700-e1a6-416b-bba1-5c211439c9d9\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.172862 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d79c5700-e1a6-416b-bba1-5c211439c9d9-run-httpd\") pod \"d79c5700-e1a6-416b-bba1-5c211439c9d9\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.172921 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-config-data\") pod \"d79c5700-e1a6-416b-bba1-5c211439c9d9\" (UID: \"d79c5700-e1a6-416b-bba1-5c211439c9d9\") " Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.174155 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d79c5700-e1a6-416b-bba1-5c211439c9d9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d79c5700-e1a6-416b-bba1-5c211439c9d9" (UID: "d79c5700-e1a6-416b-bba1-5c211439c9d9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.181235 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d79c5700-e1a6-416b-bba1-5c211439c9d9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d79c5700-e1a6-416b-bba1-5c211439c9d9" (UID: "d79c5700-e1a6-416b-bba1-5c211439c9d9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.189583 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-scripts" (OuterVolumeSpecName: "scripts") pod "d79c5700-e1a6-416b-bba1-5c211439c9d9" (UID: "d79c5700-e1a6-416b-bba1-5c211439c9d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.189967 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d79c5700-e1a6-416b-bba1-5c211439c9d9-kube-api-access-r669x" (OuterVolumeSpecName: "kube-api-access-r669x") pod "d79c5700-e1a6-416b-bba1-5c211439c9d9" (UID: "d79c5700-e1a6-416b-bba1-5c211439c9d9"). InnerVolumeSpecName "kube-api-access-r669x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.228504 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d79c5700-e1a6-416b-bba1-5c211439c9d9" (UID: "d79c5700-e1a6-416b-bba1-5c211439c9d9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.258790 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d79c5700-e1a6-416b-bba1-5c211439c9d9" (UID: "d79c5700-e1a6-416b-bba1-5c211439c9d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.275340 4739 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.275365 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.275375 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r669x\" (UniqueName: \"kubernetes.io/projected/d79c5700-e1a6-416b-bba1-5c211439c9d9-kube-api-access-r669x\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.275385 4739 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d79c5700-e1a6-416b-bba1-5c211439c9d9-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.275393 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.275400 4739 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d79c5700-e1a6-416b-bba1-5c211439c9d9-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.278378 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-config-data" (OuterVolumeSpecName: "config-data") pod "d79c5700-e1a6-416b-bba1-5c211439c9d9" (UID: "d79c5700-e1a6-416b-bba1-5c211439c9d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.376735 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d79c5700-e1a6-416b-bba1-5c211439c9d9-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.938198 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d79c5700-e1a6-416b-bba1-5c211439c9d9","Type":"ContainerDied","Data":"38bfbd7e528d0885e7d8a70130838f8fb684d4506bfd857ad5ef478a4cd70aca"} Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.938516 4739 scope.go:117] "RemoveContainer" containerID="3d096b0f728cd7c50a87d1e77a88cd58239e705a29aaa6fa8ee016ed1cfdb9a9" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.938229 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.963205 4739 scope.go:117] "RemoveContainer" containerID="d3bb11e2c8f408e8499881b7bb9857a7ad272ecb71d08fcc893c02aeb5afb5b4" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.972137 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.982980 4739 scope.go:117] "RemoveContainer" containerID="d34f390505dfaf24ddeea1b458f02ef3eb977f5d31823a8f72cd3ef73d02b41f" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.985702 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.999234 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:54:05 crc kubenswrapper[4739]: E1201 15:54:05.999614 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="sg-core" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.999637 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="sg-core" Dec 01 15:54:05 crc kubenswrapper[4739]: E1201 15:54:05.999658 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="ceilometer-notification-agent" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.999664 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="ceilometer-notification-agent" Dec 01 15:54:05 crc kubenswrapper[4739]: E1201 15:54:05.999677 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="ceilometer-central-agent" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.999685 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="ceilometer-central-agent" Dec 01 15:54:05 crc kubenswrapper[4739]: E1201 15:54:05.999693 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="proxy-httpd" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.999699 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="proxy-httpd" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.999851 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="ceilometer-central-agent" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.999869 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="sg-core" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.999879 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="ceilometer-notification-agent" Dec 01 15:54:05 crc kubenswrapper[4739]: I1201 15:54:05.999892 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" containerName="proxy-httpd" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.001246 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.003905 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.004126 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.016819 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.032007 4739 scope.go:117] "RemoveContainer" containerID="caf209b4d2ee1eb81d3f05a38cdc7cf0fa764be103faeb2ea353a40594224d76" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.189088 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.189151 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b4ee7d9-baee-491b-8663-461e89028fa2-run-httpd\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.189182 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-config-data\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.189222 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-scripts\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.189273 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.189319 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b4ee7d9-baee-491b-8663-461e89028fa2-log-httpd\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.189363 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbmtn\" (UniqueName: \"kubernetes.io/projected/4b4ee7d9-baee-491b-8663-461e89028fa2-kube-api-access-jbmtn\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.294219 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbmtn\" (UniqueName: \"kubernetes.io/projected/4b4ee7d9-baee-491b-8663-461e89028fa2-kube-api-access-jbmtn\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.294359 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.294458 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b4ee7d9-baee-491b-8663-461e89028fa2-run-httpd\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.294512 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-config-data\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.294583 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-scripts\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.294668 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.294758 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b4ee7d9-baee-491b-8663-461e89028fa2-log-httpd\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.295610 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b4ee7d9-baee-491b-8663-461e89028fa2-log-httpd\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.304312 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b4ee7d9-baee-491b-8663-461e89028fa2-run-httpd\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.310279 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.310923 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-config-data\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.320326 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-scripts\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.325688 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbmtn\" (UniqueName: \"kubernetes.io/projected/4b4ee7d9-baee-491b-8663-461e89028fa2-kube-api-access-jbmtn\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.326052 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.492274 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d79c5700-e1a6-416b-bba1-5c211439c9d9" path="/var/lib/kubelet/pods/d79c5700-e1a6-416b-bba1-5c211439c9d9/volumes" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.614664 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:54:06 crc kubenswrapper[4739]: I1201 15:54:06.974152 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:54:07 crc kubenswrapper[4739]: I1201 15:54:07.105716 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:54:07 crc kubenswrapper[4739]: W1201 15:54:07.117587 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b4ee7d9_baee_491b_8663_461e89028fa2.slice/crio-2ccfcab29dbecabd502499798681b9bebacc81d0081c27995d1d6373ed54e4e1 WatchSource:0}: Error finding container 2ccfcab29dbecabd502499798681b9bebacc81d0081c27995d1d6373ed54e4e1: Status 404 returned error can't find the container with id 2ccfcab29dbecabd502499798681b9bebacc81d0081c27995d1d6373ed54e4e1 Dec 01 15:54:07 crc kubenswrapper[4739]: I1201 15:54:07.956843 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b4ee7d9-baee-491b-8663-461e89028fa2","Type":"ContainerStarted","Data":"2ccfcab29dbecabd502499798681b9bebacc81d0081c27995d1d6373ed54e4e1"} Dec 01 15:54:08 crc kubenswrapper[4739]: I1201 15:54:08.973403 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b4ee7d9-baee-491b-8663-461e89028fa2","Type":"ContainerStarted","Data":"d64c02087f02dbbb8591dee821258b2698108bde07bfdd6549346b4407cd56f8"} Dec 01 15:54:08 crc kubenswrapper[4739]: I1201 15:54:08.973734 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b4ee7d9-baee-491b-8663-461e89028fa2","Type":"ContainerStarted","Data":"be779c8cac2cdd9fa5ea82a05062e9221ae3999b7c4fbfe6b4e72cedfe35a461"} Dec 01 15:54:09 crc kubenswrapper[4739]: I1201 15:54:09.986381 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b4ee7d9-baee-491b-8663-461e89028fa2","Type":"ContainerStarted","Data":"0cc96695e1e1284653477ead72552d8628ee825acd11d60f11e0811b31143f55"} Dec 01 15:54:10 crc kubenswrapper[4739]: I1201 15:54:10.998780 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b4ee7d9-baee-491b-8663-461e89028fa2","Type":"ContainerStarted","Data":"ca3820b86fb7dfa5f667c711afb764f9439cd7ced3610355bed4b0e880bdcfd1"} Dec 01 15:54:10 crc kubenswrapper[4739]: I1201 15:54:10.998900 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="ceilometer-central-agent" containerID="cri-o://be779c8cac2cdd9fa5ea82a05062e9221ae3999b7c4fbfe6b4e72cedfe35a461" gracePeriod=30 Dec 01 15:54:10 crc kubenswrapper[4739]: I1201 15:54:10.998926 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="sg-core" containerID="cri-o://0cc96695e1e1284653477ead72552d8628ee825acd11d60f11e0811b31143f55" gracePeriod=30 Dec 01 15:54:10 crc kubenswrapper[4739]: I1201 15:54:10.998935 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 15:54:10 crc kubenswrapper[4739]: I1201 15:54:10.998986 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="ceilometer-notification-agent" containerID="cri-o://d64c02087f02dbbb8591dee821258b2698108bde07bfdd6549346b4407cd56f8" gracePeriod=30 Dec 01 15:54:10 crc kubenswrapper[4739]: I1201 15:54:10.998929 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="proxy-httpd" containerID="cri-o://ca3820b86fb7dfa5f667c711afb764f9439cd7ced3610355bed4b0e880bdcfd1" gracePeriod=30 Dec 01 15:54:11 crc kubenswrapper[4739]: I1201 15:54:11.032699 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.485133965 podStartE2EDuration="6.032676788s" podCreationTimestamp="2025-12-01 15:54:05 +0000 UTC" firstStartedPulling="2025-12-01 15:54:07.124173184 +0000 UTC m=+1148.949919318" lastFinishedPulling="2025-12-01 15:54:10.671716047 +0000 UTC m=+1152.497462141" observedRunningTime="2025-12-01 15:54:11.025745653 +0000 UTC m=+1152.851491747" watchObservedRunningTime="2025-12-01 15:54:11.032676788 +0000 UTC m=+1152.858422882" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.010763 4739 generic.go:334] "Generic (PLEG): container finished" podID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerID="0cc96695e1e1284653477ead72552d8628ee825acd11d60f11e0811b31143f55" exitCode=2 Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.011038 4739 generic.go:334] "Generic (PLEG): container finished" podID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerID="d64c02087f02dbbb8591dee821258b2698108bde07bfdd6549346b4407cd56f8" exitCode=0 Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.010845 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b4ee7d9-baee-491b-8663-461e89028fa2","Type":"ContainerDied","Data":"0cc96695e1e1284653477ead72552d8628ee825acd11d60f11e0811b31143f55"} Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.011090 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b4ee7d9-baee-491b-8663-461e89028fa2","Type":"ContainerDied","Data":"d64c02087f02dbbb8591dee821258b2698108bde07bfdd6549346b4407cd56f8"} Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.385455 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-8rbwc"] Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.386729 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8rbwc" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.400946 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8rbwc"] Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.471508 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-wfs6d"] Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.472559 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wfs6d" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.489477 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wfs6d"] Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.526277 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6z9n\" (UniqueName: \"kubernetes.io/projected/594ff868-2963-45d0-ba4f-8e96f7e2afa1-kube-api-access-b6z9n\") pod \"nova-api-db-create-8rbwc\" (UID: \"594ff868-2963-45d0-ba4f-8e96f7e2afa1\") " pod="openstack/nova-api-db-create-8rbwc" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.526346 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/594ff868-2963-45d0-ba4f-8e96f7e2afa1-operator-scripts\") pod \"nova-api-db-create-8rbwc\" (UID: \"594ff868-2963-45d0-ba4f-8e96f7e2afa1\") " pod="openstack/nova-api-db-create-8rbwc" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.526565 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drzvn\" (UniqueName: \"kubernetes.io/projected/2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15-kube-api-access-drzvn\") pod \"nova-cell0-db-create-wfs6d\" (UID: \"2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15\") " pod="openstack/nova-cell0-db-create-wfs6d" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.526671 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15-operator-scripts\") pod \"nova-cell0-db-create-wfs6d\" (UID: \"2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15\") " pod="openstack/nova-cell0-db-create-wfs6d" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.581955 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-xtpmt"] Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.583083 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xtpmt" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.620304 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-d474-account-create-update-g7qz8"] Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.622826 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d474-account-create-update-g7qz8" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.625324 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.637293 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6zgf\" (UniqueName: \"kubernetes.io/projected/c795dbe9-11c4-4ae2-a20f-63e964a8665c-kube-api-access-q6zgf\") pod \"nova-cell1-db-create-xtpmt\" (UID: \"c795dbe9-11c4-4ae2-a20f-63e964a8665c\") " pod="openstack/nova-cell1-db-create-xtpmt" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.637392 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/594ff868-2963-45d0-ba4f-8e96f7e2afa1-operator-scripts\") pod \"nova-api-db-create-8rbwc\" (UID: \"594ff868-2963-45d0-ba4f-8e96f7e2afa1\") " pod="openstack/nova-api-db-create-8rbwc" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.637448 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d7f9b4c-d167-4b93-9a74-abb347111f76-operator-scripts\") pod \"nova-api-d474-account-create-update-g7qz8\" (UID: \"4d7f9b4c-d167-4b93-9a74-abb347111f76\") " pod="openstack/nova-api-d474-account-create-update-g7qz8" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.637511 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drzvn\" (UniqueName: \"kubernetes.io/projected/2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15-kube-api-access-drzvn\") pod \"nova-cell0-db-create-wfs6d\" (UID: \"2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15\") " pod="openstack/nova-cell0-db-create-wfs6d" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.637572 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15-operator-scripts\") pod \"nova-cell0-db-create-wfs6d\" (UID: \"2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15\") " pod="openstack/nova-cell0-db-create-wfs6d" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.637696 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsnk8\" (UniqueName: \"kubernetes.io/projected/4d7f9b4c-d167-4b93-9a74-abb347111f76-kube-api-access-xsnk8\") pod \"nova-api-d474-account-create-update-g7qz8\" (UID: \"4d7f9b4c-d167-4b93-9a74-abb347111f76\") " pod="openstack/nova-api-d474-account-create-update-g7qz8" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.637717 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c795dbe9-11c4-4ae2-a20f-63e964a8665c-operator-scripts\") pod \"nova-cell1-db-create-xtpmt\" (UID: \"c795dbe9-11c4-4ae2-a20f-63e964a8665c\") " pod="openstack/nova-cell1-db-create-xtpmt" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.637768 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6z9n\" (UniqueName: \"kubernetes.io/projected/594ff868-2963-45d0-ba4f-8e96f7e2afa1-kube-api-access-b6z9n\") pod \"nova-api-db-create-8rbwc\" (UID: \"594ff868-2963-45d0-ba4f-8e96f7e2afa1\") " pod="openstack/nova-api-db-create-8rbwc" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.639394 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/594ff868-2963-45d0-ba4f-8e96f7e2afa1-operator-scripts\") pod \"nova-api-db-create-8rbwc\" (UID: \"594ff868-2963-45d0-ba4f-8e96f7e2afa1\") " pod="openstack/nova-api-db-create-8rbwc" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.639642 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15-operator-scripts\") pod \"nova-cell0-db-create-wfs6d\" (UID: \"2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15\") " pod="openstack/nova-cell0-db-create-wfs6d" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.645479 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xtpmt"] Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.655141 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-d474-account-create-update-g7qz8"] Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.664229 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6z9n\" (UniqueName: \"kubernetes.io/projected/594ff868-2963-45d0-ba4f-8e96f7e2afa1-kube-api-access-b6z9n\") pod \"nova-api-db-create-8rbwc\" (UID: \"594ff868-2963-45d0-ba4f-8e96f7e2afa1\") " pod="openstack/nova-api-db-create-8rbwc" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.671968 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drzvn\" (UniqueName: \"kubernetes.io/projected/2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15-kube-api-access-drzvn\") pod \"nova-cell0-db-create-wfs6d\" (UID: \"2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15\") " pod="openstack/nova-cell0-db-create-wfs6d" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.704407 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8rbwc" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.739122 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsnk8\" (UniqueName: \"kubernetes.io/projected/4d7f9b4c-d167-4b93-9a74-abb347111f76-kube-api-access-xsnk8\") pod \"nova-api-d474-account-create-update-g7qz8\" (UID: \"4d7f9b4c-d167-4b93-9a74-abb347111f76\") " pod="openstack/nova-api-d474-account-create-update-g7qz8" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.739378 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c795dbe9-11c4-4ae2-a20f-63e964a8665c-operator-scripts\") pod \"nova-cell1-db-create-xtpmt\" (UID: \"c795dbe9-11c4-4ae2-a20f-63e964a8665c\") " pod="openstack/nova-cell1-db-create-xtpmt" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.739427 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6zgf\" (UniqueName: \"kubernetes.io/projected/c795dbe9-11c4-4ae2-a20f-63e964a8665c-kube-api-access-q6zgf\") pod \"nova-cell1-db-create-xtpmt\" (UID: \"c795dbe9-11c4-4ae2-a20f-63e964a8665c\") " pod="openstack/nova-cell1-db-create-xtpmt" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.739461 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d7f9b4c-d167-4b93-9a74-abb347111f76-operator-scripts\") pod \"nova-api-d474-account-create-update-g7qz8\" (UID: \"4d7f9b4c-d167-4b93-9a74-abb347111f76\") " pod="openstack/nova-api-d474-account-create-update-g7qz8" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.740059 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d7f9b4c-d167-4b93-9a74-abb347111f76-operator-scripts\") pod \"nova-api-d474-account-create-update-g7qz8\" (UID: \"4d7f9b4c-d167-4b93-9a74-abb347111f76\") " pod="openstack/nova-api-d474-account-create-update-g7qz8" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.740817 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c795dbe9-11c4-4ae2-a20f-63e964a8665c-operator-scripts\") pod \"nova-cell1-db-create-xtpmt\" (UID: \"c795dbe9-11c4-4ae2-a20f-63e964a8665c\") " pod="openstack/nova-cell1-db-create-xtpmt" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.758360 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6zgf\" (UniqueName: \"kubernetes.io/projected/c795dbe9-11c4-4ae2-a20f-63e964a8665c-kube-api-access-q6zgf\") pod \"nova-cell1-db-create-xtpmt\" (UID: \"c795dbe9-11c4-4ae2-a20f-63e964a8665c\") " pod="openstack/nova-cell1-db-create-xtpmt" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.763374 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsnk8\" (UniqueName: \"kubernetes.io/projected/4d7f9b4c-d167-4b93-9a74-abb347111f76-kube-api-access-xsnk8\") pod \"nova-api-d474-account-create-update-g7qz8\" (UID: \"4d7f9b4c-d167-4b93-9a74-abb347111f76\") " pod="openstack/nova-api-d474-account-create-update-g7qz8" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.793841 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-8179-account-create-update-57dw7"] Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.794875 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8179-account-create-update-57dw7" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.797120 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.806815 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8179-account-create-update-57dw7"] Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.833151 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wfs6d" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.842120 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq2dp\" (UniqueName: \"kubernetes.io/projected/f400ecb1-7790-4325-a68d-a9ae8e8527d1-kube-api-access-bq2dp\") pod \"nova-cell0-8179-account-create-update-57dw7\" (UID: \"f400ecb1-7790-4325-a68d-a9ae8e8527d1\") " pod="openstack/nova-cell0-8179-account-create-update-57dw7" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.842238 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f400ecb1-7790-4325-a68d-a9ae8e8527d1-operator-scripts\") pod \"nova-cell0-8179-account-create-update-57dw7\" (UID: \"f400ecb1-7790-4325-a68d-a9ae8e8527d1\") " pod="openstack/nova-cell0-8179-account-create-update-57dw7" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.925745 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xtpmt" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.943487 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f400ecb1-7790-4325-a68d-a9ae8e8527d1-operator-scripts\") pod \"nova-cell0-8179-account-create-update-57dw7\" (UID: \"f400ecb1-7790-4325-a68d-a9ae8e8527d1\") " pod="openstack/nova-cell0-8179-account-create-update-57dw7" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.943594 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq2dp\" (UniqueName: \"kubernetes.io/projected/f400ecb1-7790-4325-a68d-a9ae8e8527d1-kube-api-access-bq2dp\") pod \"nova-cell0-8179-account-create-update-57dw7\" (UID: \"f400ecb1-7790-4325-a68d-a9ae8e8527d1\") " pod="openstack/nova-cell0-8179-account-create-update-57dw7" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.944298 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f400ecb1-7790-4325-a68d-a9ae8e8527d1-operator-scripts\") pod \"nova-cell0-8179-account-create-update-57dw7\" (UID: \"f400ecb1-7790-4325-a68d-a9ae8e8527d1\") " pod="openstack/nova-cell0-8179-account-create-update-57dw7" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.946684 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d474-account-create-update-g7qz8" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.969652 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq2dp\" (UniqueName: \"kubernetes.io/projected/f400ecb1-7790-4325-a68d-a9ae8e8527d1-kube-api-access-bq2dp\") pod \"nova-cell0-8179-account-create-update-57dw7\" (UID: \"f400ecb1-7790-4325-a68d-a9ae8e8527d1\") " pod="openstack/nova-cell0-8179-account-create-update-57dw7" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.994513 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-5638-account-create-update-tcl6s"] Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.996134 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5638-account-create-update-tcl6s" Dec 01 15:54:12 crc kubenswrapper[4739]: I1201 15:54:12.997986 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 01 15:54:13 crc kubenswrapper[4739]: I1201 15:54:13.004214 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5638-account-create-update-tcl6s"] Dec 01 15:54:13 crc kubenswrapper[4739]: I1201 15:54:13.045974 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8aa61fba-e9af-4666-a96e-498c56ddeaf0-operator-scripts\") pod \"nova-cell1-5638-account-create-update-tcl6s\" (UID: \"8aa61fba-e9af-4666-a96e-498c56ddeaf0\") " pod="openstack/nova-cell1-5638-account-create-update-tcl6s" Dec 01 15:54:13 crc kubenswrapper[4739]: I1201 15:54:13.046057 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc7kt\" (UniqueName: \"kubernetes.io/projected/8aa61fba-e9af-4666-a96e-498c56ddeaf0-kube-api-access-dc7kt\") pod \"nova-cell1-5638-account-create-update-tcl6s\" (UID: \"8aa61fba-e9af-4666-a96e-498c56ddeaf0\") " pod="openstack/nova-cell1-5638-account-create-update-tcl6s" Dec 01 15:54:13 crc kubenswrapper[4739]: I1201 15:54:13.141564 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8179-account-create-update-57dw7" Dec 01 15:54:13 crc kubenswrapper[4739]: I1201 15:54:13.147849 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc7kt\" (UniqueName: \"kubernetes.io/projected/8aa61fba-e9af-4666-a96e-498c56ddeaf0-kube-api-access-dc7kt\") pod \"nova-cell1-5638-account-create-update-tcl6s\" (UID: \"8aa61fba-e9af-4666-a96e-498c56ddeaf0\") " pod="openstack/nova-cell1-5638-account-create-update-tcl6s" Dec 01 15:54:13 crc kubenswrapper[4739]: I1201 15:54:13.148013 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8aa61fba-e9af-4666-a96e-498c56ddeaf0-operator-scripts\") pod \"nova-cell1-5638-account-create-update-tcl6s\" (UID: \"8aa61fba-e9af-4666-a96e-498c56ddeaf0\") " pod="openstack/nova-cell1-5638-account-create-update-tcl6s" Dec 01 15:54:13 crc kubenswrapper[4739]: I1201 15:54:13.148949 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8aa61fba-e9af-4666-a96e-498c56ddeaf0-operator-scripts\") pod \"nova-cell1-5638-account-create-update-tcl6s\" (UID: \"8aa61fba-e9af-4666-a96e-498c56ddeaf0\") " pod="openstack/nova-cell1-5638-account-create-update-tcl6s" Dec 01 15:54:13 crc kubenswrapper[4739]: I1201 15:54:13.163494 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc7kt\" (UniqueName: \"kubernetes.io/projected/8aa61fba-e9af-4666-a96e-498c56ddeaf0-kube-api-access-dc7kt\") pod \"nova-cell1-5638-account-create-update-tcl6s\" (UID: \"8aa61fba-e9af-4666-a96e-498c56ddeaf0\") " pod="openstack/nova-cell1-5638-account-create-update-tcl6s" Dec 01 15:54:13 crc kubenswrapper[4739]: I1201 15:54:13.183395 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8rbwc"] Dec 01 15:54:13 crc kubenswrapper[4739]: W1201 15:54:13.187851 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod594ff868_2963_45d0_ba4f_8e96f7e2afa1.slice/crio-04f8392b5f73ed5160b353f3adb55dc253e497992861ad5d78f8f726c2f505eb WatchSource:0}: Error finding container 04f8392b5f73ed5160b353f3adb55dc253e497992861ad5d78f8f726c2f505eb: Status 404 returned error can't find the container with id 04f8392b5f73ed5160b353f3adb55dc253e497992861ad5d78f8f726c2f505eb Dec 01 15:54:13 crc kubenswrapper[4739]: I1201 15:54:13.328120 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5638-account-create-update-tcl6s" Dec 01 15:54:13 crc kubenswrapper[4739]: I1201 15:54:13.329139 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wfs6d"] Dec 01 15:54:13 crc kubenswrapper[4739]: I1201 15:54:13.409228 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xtpmt"] Dec 01 15:54:13 crc kubenswrapper[4739]: W1201 15:54:13.426624 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc795dbe9_11c4_4ae2_a20f_63e964a8665c.slice/crio-164b950e8a10274fb2688ca8a49cccde6714b14adeec62dbfad60a7aff92e4be WatchSource:0}: Error finding container 164b950e8a10274fb2688ca8a49cccde6714b14adeec62dbfad60a7aff92e4be: Status 404 returned error can't find the container with id 164b950e8a10274fb2688ca8a49cccde6714b14adeec62dbfad60a7aff92e4be Dec 01 15:54:13 crc kubenswrapper[4739]: I1201 15:54:13.494179 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-d474-account-create-update-g7qz8"] Dec 01 15:54:13 crc kubenswrapper[4739]: W1201 15:54:13.514660 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d7f9b4c_d167_4b93_9a74_abb347111f76.slice/crio-f082ea1054e8a0daa94184c1262ef4baae62c7dcff9c7eb92fd901fddc33cdcd WatchSource:0}: Error finding container f082ea1054e8a0daa94184c1262ef4baae62c7dcff9c7eb92fd901fddc33cdcd: Status 404 returned error can't find the container with id f082ea1054e8a0daa94184c1262ef4baae62c7dcff9c7eb92fd901fddc33cdcd Dec 01 15:54:13 crc kubenswrapper[4739]: W1201 15:54:13.621888 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8aa61fba_e9af_4666_a96e_498c56ddeaf0.slice/crio-319ae366fdb5fcf1360079aed09ccd2c9de6acac974951bea80f30c1c56674b2 WatchSource:0}: Error finding container 319ae366fdb5fcf1360079aed09ccd2c9de6acac974951bea80f30c1c56674b2: Status 404 returned error can't find the container with id 319ae366fdb5fcf1360079aed09ccd2c9de6acac974951bea80f30c1c56674b2 Dec 01 15:54:13 crc kubenswrapper[4739]: I1201 15:54:13.622562 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5638-account-create-update-tcl6s"] Dec 01 15:54:13 crc kubenswrapper[4739]: I1201 15:54:13.729127 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8179-account-create-update-57dw7"] Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.039821 4739 generic.go:334] "Generic (PLEG): container finished" podID="2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15" containerID="3f6ef0a436a192023393027b9011e2de3fbc772e6a45cf037ee55a136e52ab94" exitCode=0 Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.039911 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wfs6d" event={"ID":"2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15","Type":"ContainerDied","Data":"3f6ef0a436a192023393027b9011e2de3fbc772e6a45cf037ee55a136e52ab94"} Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.039944 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wfs6d" event={"ID":"2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15","Type":"ContainerStarted","Data":"0c7979e5b61abe50628be6dab3e6d0613dcbfef9eea5622c320fa1b9325fd058"} Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.041808 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5638-account-create-update-tcl6s" event={"ID":"8aa61fba-e9af-4666-a96e-498c56ddeaf0","Type":"ContainerStarted","Data":"67d173165ff5bf022fd97d4534a03846fdf75b53cc059c6e8843f3c61276236b"} Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.041848 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5638-account-create-update-tcl6s" event={"ID":"8aa61fba-e9af-4666-a96e-498c56ddeaf0","Type":"ContainerStarted","Data":"319ae366fdb5fcf1360079aed09ccd2c9de6acac974951bea80f30c1c56674b2"} Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.043667 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8179-account-create-update-57dw7" event={"ID":"f400ecb1-7790-4325-a68d-a9ae8e8527d1","Type":"ContainerStarted","Data":"8e76b1a4598841709a798c7d605f6b410086ae6f2e45802299e93ac52f40fe9d"} Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.043710 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8179-account-create-update-57dw7" event={"ID":"f400ecb1-7790-4325-a68d-a9ae8e8527d1","Type":"ContainerStarted","Data":"169af48650179c35bc9cffbf6a22b518fc2e6fc23a1949c4a877734f3e85df4d"} Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.044970 4739 generic.go:334] "Generic (PLEG): container finished" podID="c795dbe9-11c4-4ae2-a20f-63e964a8665c" containerID="e9fe065410204039b095972e576064aef9d5b8a20668e67d7eaf2fa1ff8f426d" exitCode=0 Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.045041 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xtpmt" event={"ID":"c795dbe9-11c4-4ae2-a20f-63e964a8665c","Type":"ContainerDied","Data":"e9fe065410204039b095972e576064aef9d5b8a20668e67d7eaf2fa1ff8f426d"} Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.045063 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xtpmt" event={"ID":"c795dbe9-11c4-4ae2-a20f-63e964a8665c","Type":"ContainerStarted","Data":"164b950e8a10274fb2688ca8a49cccde6714b14adeec62dbfad60a7aff92e4be"} Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.046432 4739 generic.go:334] "Generic (PLEG): container finished" podID="594ff868-2963-45d0-ba4f-8e96f7e2afa1" containerID="0b19591959da1c430cf8a8d6003adce34472fd08a3f8f479563df60f8174d8bb" exitCode=0 Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.046486 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8rbwc" event={"ID":"594ff868-2963-45d0-ba4f-8e96f7e2afa1","Type":"ContainerDied","Data":"0b19591959da1c430cf8a8d6003adce34472fd08a3f8f479563df60f8174d8bb"} Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.046505 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8rbwc" event={"ID":"594ff868-2963-45d0-ba4f-8e96f7e2afa1","Type":"ContainerStarted","Data":"04f8392b5f73ed5160b353f3adb55dc253e497992861ad5d78f8f726c2f505eb"} Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.047720 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d474-account-create-update-g7qz8" event={"ID":"4d7f9b4c-d167-4b93-9a74-abb347111f76","Type":"ContainerStarted","Data":"1be628c6fd28e82ded158029d635c2a2e5ef07c2e4fee3bc7ad33c8fc016442d"} Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.047740 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d474-account-create-update-g7qz8" event={"ID":"4d7f9b4c-d167-4b93-9a74-abb347111f76","Type":"ContainerStarted","Data":"f082ea1054e8a0daa94184c1262ef4baae62c7dcff9c7eb92fd901fddc33cdcd"} Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.072541 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-5638-account-create-update-tcl6s" podStartSLOduration=2.072521856 podStartE2EDuration="2.072521856s" podCreationTimestamp="2025-12-01 15:54:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:54:14.071329939 +0000 UTC m=+1155.897076053" watchObservedRunningTime="2025-12-01 15:54:14.072521856 +0000 UTC m=+1155.898267950" Dec 01 15:54:14 crc kubenswrapper[4739]: I1201 15:54:14.089841 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-8179-account-create-update-57dw7" podStartSLOduration=2.089817951 podStartE2EDuration="2.089817951s" podCreationTimestamp="2025-12-01 15:54:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:54:14.086016013 +0000 UTC m=+1155.911762127" watchObservedRunningTime="2025-12-01 15:54:14.089817951 +0000 UTC m=+1155.915564055" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.073590 4739 generic.go:334] "Generic (PLEG): container finished" podID="8aa61fba-e9af-4666-a96e-498c56ddeaf0" containerID="67d173165ff5bf022fd97d4534a03846fdf75b53cc059c6e8843f3c61276236b" exitCode=0 Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.073779 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5638-account-create-update-tcl6s" event={"ID":"8aa61fba-e9af-4666-a96e-498c56ddeaf0","Type":"ContainerDied","Data":"67d173165ff5bf022fd97d4534a03846fdf75b53cc059c6e8843f3c61276236b"} Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.077908 4739 generic.go:334] "Generic (PLEG): container finished" podID="f400ecb1-7790-4325-a68d-a9ae8e8527d1" containerID="8e76b1a4598841709a798c7d605f6b410086ae6f2e45802299e93ac52f40fe9d" exitCode=0 Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.077967 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8179-account-create-update-57dw7" event={"ID":"f400ecb1-7790-4325-a68d-a9ae8e8527d1","Type":"ContainerDied","Data":"8e76b1a4598841709a798c7d605f6b410086ae6f2e45802299e93ac52f40fe9d"} Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.080549 4739 generic.go:334] "Generic (PLEG): container finished" podID="4d7f9b4c-d167-4b93-9a74-abb347111f76" containerID="1be628c6fd28e82ded158029d635c2a2e5ef07c2e4fee3bc7ad33c8fc016442d" exitCode=0 Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.081665 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d474-account-create-update-g7qz8" event={"ID":"4d7f9b4c-d167-4b93-9a74-abb347111f76","Type":"ContainerDied","Data":"1be628c6fd28e82ded158029d635c2a2e5ef07c2e4fee3bc7ad33c8fc016442d"} Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.489624 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wfs6d" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.593103 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drzvn\" (UniqueName: \"kubernetes.io/projected/2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15-kube-api-access-drzvn\") pod \"2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15\" (UID: \"2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15\") " Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.593324 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15-operator-scripts\") pod \"2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15\" (UID: \"2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15\") " Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.594120 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15" (UID: "2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.604549 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15-kube-api-access-drzvn" (OuterVolumeSpecName: "kube-api-access-drzvn") pod "2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15" (UID: "2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15"). InnerVolumeSpecName "kube-api-access-drzvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.644330 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8rbwc" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.658936 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d474-account-create-update-g7qz8" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.668928 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xtpmt" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.694799 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6z9n\" (UniqueName: \"kubernetes.io/projected/594ff868-2963-45d0-ba4f-8e96f7e2afa1-kube-api-access-b6z9n\") pod \"594ff868-2963-45d0-ba4f-8e96f7e2afa1\" (UID: \"594ff868-2963-45d0-ba4f-8e96f7e2afa1\") " Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.695053 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/594ff868-2963-45d0-ba4f-8e96f7e2afa1-operator-scripts\") pod \"594ff868-2963-45d0-ba4f-8e96f7e2afa1\" (UID: \"594ff868-2963-45d0-ba4f-8e96f7e2afa1\") " Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.695560 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.695578 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drzvn\" (UniqueName: \"kubernetes.io/projected/2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15-kube-api-access-drzvn\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.697006 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/594ff868-2963-45d0-ba4f-8e96f7e2afa1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "594ff868-2963-45d0-ba4f-8e96f7e2afa1" (UID: "594ff868-2963-45d0-ba4f-8e96f7e2afa1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.700359 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/594ff868-2963-45d0-ba4f-8e96f7e2afa1-kube-api-access-b6z9n" (OuterVolumeSpecName: "kube-api-access-b6z9n") pod "594ff868-2963-45d0-ba4f-8e96f7e2afa1" (UID: "594ff868-2963-45d0-ba4f-8e96f7e2afa1"). InnerVolumeSpecName "kube-api-access-b6z9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.796742 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsnk8\" (UniqueName: \"kubernetes.io/projected/4d7f9b4c-d167-4b93-9a74-abb347111f76-kube-api-access-xsnk8\") pod \"4d7f9b4c-d167-4b93-9a74-abb347111f76\" (UID: \"4d7f9b4c-d167-4b93-9a74-abb347111f76\") " Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.796841 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d7f9b4c-d167-4b93-9a74-abb347111f76-operator-scripts\") pod \"4d7f9b4c-d167-4b93-9a74-abb347111f76\" (UID: \"4d7f9b4c-d167-4b93-9a74-abb347111f76\") " Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.796859 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c795dbe9-11c4-4ae2-a20f-63e964a8665c-operator-scripts\") pod \"c795dbe9-11c4-4ae2-a20f-63e964a8665c\" (UID: \"c795dbe9-11c4-4ae2-a20f-63e964a8665c\") " Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.796996 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6zgf\" (UniqueName: \"kubernetes.io/projected/c795dbe9-11c4-4ae2-a20f-63e964a8665c-kube-api-access-q6zgf\") pod \"c795dbe9-11c4-4ae2-a20f-63e964a8665c\" (UID: \"c795dbe9-11c4-4ae2-a20f-63e964a8665c\") " Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.797309 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6z9n\" (UniqueName: \"kubernetes.io/projected/594ff868-2963-45d0-ba4f-8e96f7e2afa1-kube-api-access-b6z9n\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.797326 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/594ff868-2963-45d0-ba4f-8e96f7e2afa1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.797889 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d7f9b4c-d167-4b93-9a74-abb347111f76-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4d7f9b4c-d167-4b93-9a74-abb347111f76" (UID: "4d7f9b4c-d167-4b93-9a74-abb347111f76"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.798305 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c795dbe9-11c4-4ae2-a20f-63e964a8665c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c795dbe9-11c4-4ae2-a20f-63e964a8665c" (UID: "c795dbe9-11c4-4ae2-a20f-63e964a8665c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.800742 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d7f9b4c-d167-4b93-9a74-abb347111f76-kube-api-access-xsnk8" (OuterVolumeSpecName: "kube-api-access-xsnk8") pod "4d7f9b4c-d167-4b93-9a74-abb347111f76" (UID: "4d7f9b4c-d167-4b93-9a74-abb347111f76"). InnerVolumeSpecName "kube-api-access-xsnk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.800872 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c795dbe9-11c4-4ae2-a20f-63e964a8665c-kube-api-access-q6zgf" (OuterVolumeSpecName: "kube-api-access-q6zgf") pod "c795dbe9-11c4-4ae2-a20f-63e964a8665c" (UID: "c795dbe9-11c4-4ae2-a20f-63e964a8665c"). InnerVolumeSpecName "kube-api-access-q6zgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.898924 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6zgf\" (UniqueName: \"kubernetes.io/projected/c795dbe9-11c4-4ae2-a20f-63e964a8665c-kube-api-access-q6zgf\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.898955 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsnk8\" (UniqueName: \"kubernetes.io/projected/4d7f9b4c-d167-4b93-9a74-abb347111f76-kube-api-access-xsnk8\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.898965 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d7f9b4c-d167-4b93-9a74-abb347111f76-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:15 crc kubenswrapper[4739]: I1201 15:54:15.898974 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c795dbe9-11c4-4ae2-a20f-63e964a8665c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.091982 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8rbwc" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.092669 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8rbwc" event={"ID":"594ff868-2963-45d0-ba4f-8e96f7e2afa1","Type":"ContainerDied","Data":"04f8392b5f73ed5160b353f3adb55dc253e497992861ad5d78f8f726c2f505eb"} Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.092749 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04f8392b5f73ed5160b353f3adb55dc253e497992861ad5d78f8f726c2f505eb" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.094133 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d474-account-create-update-g7qz8" event={"ID":"4d7f9b4c-d167-4b93-9a74-abb347111f76","Type":"ContainerDied","Data":"f082ea1054e8a0daa94184c1262ef4baae62c7dcff9c7eb92fd901fddc33cdcd"} Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.094170 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f082ea1054e8a0daa94184c1262ef4baae62c7dcff9c7eb92fd901fddc33cdcd" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.094222 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d474-account-create-update-g7qz8" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.097130 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wfs6d" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.097147 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wfs6d" event={"ID":"2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15","Type":"ContainerDied","Data":"0c7979e5b61abe50628be6dab3e6d0613dcbfef9eea5622c320fa1b9325fd058"} Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.097184 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c7979e5b61abe50628be6dab3e6d0613dcbfef9eea5622c320fa1b9325fd058" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.098855 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xtpmt" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.098901 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xtpmt" event={"ID":"c795dbe9-11c4-4ae2-a20f-63e964a8665c","Type":"ContainerDied","Data":"164b950e8a10274fb2688ca8a49cccde6714b14adeec62dbfad60a7aff92e4be"} Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.098945 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="164b950e8a10274fb2688ca8a49cccde6714b14adeec62dbfad60a7aff92e4be" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.529639 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5638-account-create-update-tcl6s" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.536969 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8179-account-create-update-57dw7" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.610442 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc7kt\" (UniqueName: \"kubernetes.io/projected/8aa61fba-e9af-4666-a96e-498c56ddeaf0-kube-api-access-dc7kt\") pod \"8aa61fba-e9af-4666-a96e-498c56ddeaf0\" (UID: \"8aa61fba-e9af-4666-a96e-498c56ddeaf0\") " Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.610512 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bq2dp\" (UniqueName: \"kubernetes.io/projected/f400ecb1-7790-4325-a68d-a9ae8e8527d1-kube-api-access-bq2dp\") pod \"f400ecb1-7790-4325-a68d-a9ae8e8527d1\" (UID: \"f400ecb1-7790-4325-a68d-a9ae8e8527d1\") " Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.610634 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8aa61fba-e9af-4666-a96e-498c56ddeaf0-operator-scripts\") pod \"8aa61fba-e9af-4666-a96e-498c56ddeaf0\" (UID: \"8aa61fba-e9af-4666-a96e-498c56ddeaf0\") " Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.610759 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f400ecb1-7790-4325-a68d-a9ae8e8527d1-operator-scripts\") pod \"f400ecb1-7790-4325-a68d-a9ae8e8527d1\" (UID: \"f400ecb1-7790-4325-a68d-a9ae8e8527d1\") " Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.611705 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f400ecb1-7790-4325-a68d-a9ae8e8527d1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f400ecb1-7790-4325-a68d-a9ae8e8527d1" (UID: "f400ecb1-7790-4325-a68d-a9ae8e8527d1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.611811 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa61fba-e9af-4666-a96e-498c56ddeaf0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8aa61fba-e9af-4666-a96e-498c56ddeaf0" (UID: "8aa61fba-e9af-4666-a96e-498c56ddeaf0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.617613 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f400ecb1-7790-4325-a68d-a9ae8e8527d1-kube-api-access-bq2dp" (OuterVolumeSpecName: "kube-api-access-bq2dp") pod "f400ecb1-7790-4325-a68d-a9ae8e8527d1" (UID: "f400ecb1-7790-4325-a68d-a9ae8e8527d1"). InnerVolumeSpecName "kube-api-access-bq2dp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.617857 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aa61fba-e9af-4666-a96e-498c56ddeaf0-kube-api-access-dc7kt" (OuterVolumeSpecName: "kube-api-access-dc7kt") pod "8aa61fba-e9af-4666-a96e-498c56ddeaf0" (UID: "8aa61fba-e9af-4666-a96e-498c56ddeaf0"). InnerVolumeSpecName "kube-api-access-dc7kt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.713972 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc7kt\" (UniqueName: \"kubernetes.io/projected/8aa61fba-e9af-4666-a96e-498c56ddeaf0-kube-api-access-dc7kt\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.714000 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bq2dp\" (UniqueName: \"kubernetes.io/projected/f400ecb1-7790-4325-a68d-a9ae8e8527d1-kube-api-access-bq2dp\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.714013 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8aa61fba-e9af-4666-a96e-498c56ddeaf0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:16 crc kubenswrapper[4739]: I1201 15:54:16.714021 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f400ecb1-7790-4325-a68d-a9ae8e8527d1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.116794 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8179-account-create-update-57dw7" event={"ID":"f400ecb1-7790-4325-a68d-a9ae8e8527d1","Type":"ContainerDied","Data":"169af48650179c35bc9cffbf6a22b518fc2e6fc23a1949c4a877734f3e85df4d"} Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.116851 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="169af48650179c35bc9cffbf6a22b518fc2e6fc23a1949c4a877734f3e85df4d" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.117587 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8179-account-create-update-57dw7" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.125316 4739 generic.go:334] "Generic (PLEG): container finished" podID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerID="be779c8cac2cdd9fa5ea82a05062e9221ae3999b7c4fbfe6b4e72cedfe35a461" exitCode=0 Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.125386 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b4ee7d9-baee-491b-8663-461e89028fa2","Type":"ContainerDied","Data":"be779c8cac2cdd9fa5ea82a05062e9221ae3999b7c4fbfe6b4e72cedfe35a461"} Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.127241 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5638-account-create-update-tcl6s" event={"ID":"8aa61fba-e9af-4666-a96e-498c56ddeaf0","Type":"ContainerDied","Data":"319ae366fdb5fcf1360079aed09ccd2c9de6acac974951bea80f30c1c56674b2"} Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.127269 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="319ae366fdb5fcf1360079aed09ccd2c9de6acac974951bea80f30c1c56674b2" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.127337 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5638-account-create-update-tcl6s" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.988230 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-626sv"] Dec 01 15:54:17 crc kubenswrapper[4739]: E1201 15:54:17.990689 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f400ecb1-7790-4325-a68d-a9ae8e8527d1" containerName="mariadb-account-create-update" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.990709 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f400ecb1-7790-4325-a68d-a9ae8e8527d1" containerName="mariadb-account-create-update" Dec 01 15:54:17 crc kubenswrapper[4739]: E1201 15:54:17.990721 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c795dbe9-11c4-4ae2-a20f-63e964a8665c" containerName="mariadb-database-create" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.990839 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="c795dbe9-11c4-4ae2-a20f-63e964a8665c" containerName="mariadb-database-create" Dec 01 15:54:17 crc kubenswrapper[4739]: E1201 15:54:17.990856 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="594ff868-2963-45d0-ba4f-8e96f7e2afa1" containerName="mariadb-database-create" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.990862 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="594ff868-2963-45d0-ba4f-8e96f7e2afa1" containerName="mariadb-database-create" Dec 01 15:54:17 crc kubenswrapper[4739]: E1201 15:54:17.990882 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15" containerName="mariadb-database-create" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.990888 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15" containerName="mariadb-database-create" Dec 01 15:54:17 crc kubenswrapper[4739]: E1201 15:54:17.990910 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d7f9b4c-d167-4b93-9a74-abb347111f76" containerName="mariadb-account-create-update" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.990916 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d7f9b4c-d167-4b93-9a74-abb347111f76" containerName="mariadb-account-create-update" Dec 01 15:54:17 crc kubenswrapper[4739]: E1201 15:54:17.990937 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aa61fba-e9af-4666-a96e-498c56ddeaf0" containerName="mariadb-account-create-update" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.990943 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aa61fba-e9af-4666-a96e-498c56ddeaf0" containerName="mariadb-account-create-update" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.991122 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15" containerName="mariadb-database-create" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.991150 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d7f9b4c-d167-4b93-9a74-abb347111f76" containerName="mariadb-account-create-update" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.991161 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="594ff868-2963-45d0-ba4f-8e96f7e2afa1" containerName="mariadb-database-create" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.991171 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f400ecb1-7790-4325-a68d-a9ae8e8527d1" containerName="mariadb-account-create-update" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.991226 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="c795dbe9-11c4-4ae2-a20f-63e964a8665c" containerName="mariadb-database-create" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.991235 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aa61fba-e9af-4666-a96e-498c56ddeaf0" containerName="mariadb-account-create-update" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.992055 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.997811 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-czskg" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.998002 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 15:54:17 crc kubenswrapper[4739]: I1201 15:54:17.998117 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 01 15:54:18 crc kubenswrapper[4739]: I1201 15:54:18.021663 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-626sv"] Dec 01 15:54:18 crc kubenswrapper[4739]: I1201 15:54:18.138884 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-scripts\") pod \"nova-cell0-conductor-db-sync-626sv\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:18 crc kubenswrapper[4739]: I1201 15:54:18.138981 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-626sv\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:18 crc kubenswrapper[4739]: I1201 15:54:18.139107 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-config-data\") pod \"nova-cell0-conductor-db-sync-626sv\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:18 crc kubenswrapper[4739]: I1201 15:54:18.139412 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsttc\" (UniqueName: \"kubernetes.io/projected/4dcf654f-3264-479d-9d65-acd256667075-kube-api-access-vsttc\") pod \"nova-cell0-conductor-db-sync-626sv\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:18 crc kubenswrapper[4739]: I1201 15:54:18.241152 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-scripts\") pod \"nova-cell0-conductor-db-sync-626sv\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:18 crc kubenswrapper[4739]: I1201 15:54:18.241263 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-626sv\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:18 crc kubenswrapper[4739]: I1201 15:54:18.241365 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-config-data\") pod \"nova-cell0-conductor-db-sync-626sv\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:18 crc kubenswrapper[4739]: I1201 15:54:18.241641 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsttc\" (UniqueName: \"kubernetes.io/projected/4dcf654f-3264-479d-9d65-acd256667075-kube-api-access-vsttc\") pod \"nova-cell0-conductor-db-sync-626sv\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:18 crc kubenswrapper[4739]: I1201 15:54:18.247087 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-config-data\") pod \"nova-cell0-conductor-db-sync-626sv\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:18 crc kubenswrapper[4739]: I1201 15:54:18.257276 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-626sv\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:18 crc kubenswrapper[4739]: I1201 15:54:18.260463 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-scripts\") pod \"nova-cell0-conductor-db-sync-626sv\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:18 crc kubenswrapper[4739]: I1201 15:54:18.262102 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsttc\" (UniqueName: \"kubernetes.io/projected/4dcf654f-3264-479d-9d65-acd256667075-kube-api-access-vsttc\") pod \"nova-cell0-conductor-db-sync-626sv\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:18 crc kubenswrapper[4739]: I1201 15:54:18.332253 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:18 crc kubenswrapper[4739]: I1201 15:54:18.805634 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-626sv"] Dec 01 15:54:19 crc kubenswrapper[4739]: I1201 15:54:19.146297 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-626sv" event={"ID":"4dcf654f-3264-479d-9d65-acd256667075","Type":"ContainerStarted","Data":"b92d28b347b3eaf574df9a1cc564c4780aed102643d9c67a9c42bbb4bd2fde5a"} Dec 01 15:54:26 crc kubenswrapper[4739]: I1201 15:54:26.222220 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-626sv" event={"ID":"4dcf654f-3264-479d-9d65-acd256667075","Type":"ContainerStarted","Data":"6f84adc8bc9432bd88b2da8732675bf16b677ebba7e9df93da652eeffb56da30"} Dec 01 15:54:26 crc kubenswrapper[4739]: I1201 15:54:26.252645 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-626sv" podStartSLOduration=2.827674817 podStartE2EDuration="9.25262735s" podCreationTimestamp="2025-12-01 15:54:17 +0000 UTC" firstStartedPulling="2025-12-01 15:54:18.811588708 +0000 UTC m=+1160.637334802" lastFinishedPulling="2025-12-01 15:54:25.236541241 +0000 UTC m=+1167.062287335" observedRunningTime="2025-12-01 15:54:26.243178109 +0000 UTC m=+1168.068924213" watchObservedRunningTime="2025-12-01 15:54:26.25262735 +0000 UTC m=+1168.078373444" Dec 01 15:54:36 crc kubenswrapper[4739]: I1201 15:54:36.328339 4739 generic.go:334] "Generic (PLEG): container finished" podID="4dcf654f-3264-479d-9d65-acd256667075" containerID="6f84adc8bc9432bd88b2da8732675bf16b677ebba7e9df93da652eeffb56da30" exitCode=0 Dec 01 15:54:36 crc kubenswrapper[4739]: I1201 15:54:36.328461 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-626sv" event={"ID":"4dcf654f-3264-479d-9d65-acd256667075","Type":"ContainerDied","Data":"6f84adc8bc9432bd88b2da8732675bf16b677ebba7e9df93da652eeffb56da30"} Dec 01 15:54:36 crc kubenswrapper[4739]: I1201 15:54:36.621207 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 01 15:54:37 crc kubenswrapper[4739]: I1201 15:54:37.752770 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:37 crc kubenswrapper[4739]: I1201 15:54:37.828328 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-config-data\") pod \"4dcf654f-3264-479d-9d65-acd256667075\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " Dec 01 15:54:37 crc kubenswrapper[4739]: I1201 15:54:37.828477 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsttc\" (UniqueName: \"kubernetes.io/projected/4dcf654f-3264-479d-9d65-acd256667075-kube-api-access-vsttc\") pod \"4dcf654f-3264-479d-9d65-acd256667075\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " Dec 01 15:54:37 crc kubenswrapper[4739]: I1201 15:54:37.828545 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-scripts\") pod \"4dcf654f-3264-479d-9d65-acd256667075\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " Dec 01 15:54:37 crc kubenswrapper[4739]: I1201 15:54:37.828748 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-combined-ca-bundle\") pod \"4dcf654f-3264-479d-9d65-acd256667075\" (UID: \"4dcf654f-3264-479d-9d65-acd256667075\") " Dec 01 15:54:37 crc kubenswrapper[4739]: I1201 15:54:37.834387 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dcf654f-3264-479d-9d65-acd256667075-kube-api-access-vsttc" (OuterVolumeSpecName: "kube-api-access-vsttc") pod "4dcf654f-3264-479d-9d65-acd256667075" (UID: "4dcf654f-3264-479d-9d65-acd256667075"). InnerVolumeSpecName "kube-api-access-vsttc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:54:37 crc kubenswrapper[4739]: I1201 15:54:37.836520 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-scripts" (OuterVolumeSpecName: "scripts") pod "4dcf654f-3264-479d-9d65-acd256667075" (UID: "4dcf654f-3264-479d-9d65-acd256667075"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:54:37 crc kubenswrapper[4739]: I1201 15:54:37.855505 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-config-data" (OuterVolumeSpecName: "config-data") pod "4dcf654f-3264-479d-9d65-acd256667075" (UID: "4dcf654f-3264-479d-9d65-acd256667075"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:54:37 crc kubenswrapper[4739]: I1201 15:54:37.874071 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4dcf654f-3264-479d-9d65-acd256667075" (UID: "4dcf654f-3264-479d-9d65-acd256667075"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:54:37 crc kubenswrapper[4739]: I1201 15:54:37.931652 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:37 crc kubenswrapper[4739]: I1201 15:54:37.931696 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:37 crc kubenswrapper[4739]: I1201 15:54:37.931709 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsttc\" (UniqueName: \"kubernetes.io/projected/4dcf654f-3264-479d-9d65-acd256667075-kube-api-access-vsttc\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:37 crc kubenswrapper[4739]: I1201 15:54:37.931723 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4dcf654f-3264-479d-9d65-acd256667075-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.353029 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-626sv" event={"ID":"4dcf654f-3264-479d-9d65-acd256667075","Type":"ContainerDied","Data":"b92d28b347b3eaf574df9a1cc564c4780aed102643d9c67a9c42bbb4bd2fde5a"} Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.353078 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b92d28b347b3eaf574df9a1cc564c4780aed102643d9c67a9c42bbb4bd2fde5a" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.353123 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-626sv" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.511274 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 15:54:38 crc kubenswrapper[4739]: E1201 15:54:38.512000 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dcf654f-3264-479d-9d65-acd256667075" containerName="nova-cell0-conductor-db-sync" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.515007 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dcf654f-3264-479d-9d65-acd256667075" containerName="nova-cell0-conductor-db-sync" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.515372 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dcf654f-3264-479d-9d65-acd256667075" containerName="nova-cell0-conductor-db-sync" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.516076 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.519303 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-czskg" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.519735 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.521894 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.643746 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50a7f5a5-c069-4e82-97e0-ecccdf443c15-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"50a7f5a5-c069-4e82-97e0-ecccdf443c15\") " pod="openstack/nova-cell0-conductor-0" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.643902 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50a7f5a5-c069-4e82-97e0-ecccdf443c15-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"50a7f5a5-c069-4e82-97e0-ecccdf443c15\") " pod="openstack/nova-cell0-conductor-0" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.644027 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rntl\" (UniqueName: \"kubernetes.io/projected/50a7f5a5-c069-4e82-97e0-ecccdf443c15-kube-api-access-7rntl\") pod \"nova-cell0-conductor-0\" (UID: \"50a7f5a5-c069-4e82-97e0-ecccdf443c15\") " pod="openstack/nova-cell0-conductor-0" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.745539 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rntl\" (UniqueName: \"kubernetes.io/projected/50a7f5a5-c069-4e82-97e0-ecccdf443c15-kube-api-access-7rntl\") pod \"nova-cell0-conductor-0\" (UID: \"50a7f5a5-c069-4e82-97e0-ecccdf443c15\") " pod="openstack/nova-cell0-conductor-0" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.745985 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50a7f5a5-c069-4e82-97e0-ecccdf443c15-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"50a7f5a5-c069-4e82-97e0-ecccdf443c15\") " pod="openstack/nova-cell0-conductor-0" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.746279 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50a7f5a5-c069-4e82-97e0-ecccdf443c15-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"50a7f5a5-c069-4e82-97e0-ecccdf443c15\") " pod="openstack/nova-cell0-conductor-0" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.753244 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50a7f5a5-c069-4e82-97e0-ecccdf443c15-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"50a7f5a5-c069-4e82-97e0-ecccdf443c15\") " pod="openstack/nova-cell0-conductor-0" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.753509 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50a7f5a5-c069-4e82-97e0-ecccdf443c15-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"50a7f5a5-c069-4e82-97e0-ecccdf443c15\") " pod="openstack/nova-cell0-conductor-0" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.765874 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rntl\" (UniqueName: \"kubernetes.io/projected/50a7f5a5-c069-4e82-97e0-ecccdf443c15-kube-api-access-7rntl\") pod \"nova-cell0-conductor-0\" (UID: \"50a7f5a5-c069-4e82-97e0-ecccdf443c15\") " pod="openstack/nova-cell0-conductor-0" Dec 01 15:54:38 crc kubenswrapper[4739]: I1201 15:54:38.882552 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 15:54:39 crc kubenswrapper[4739]: I1201 15:54:39.360223 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 15:54:39 crc kubenswrapper[4739]: I1201 15:54:39.367051 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"50a7f5a5-c069-4e82-97e0-ecccdf443c15","Type":"ContainerStarted","Data":"550a03473edd12c68fdd69efa39bb7aa7d5b4324b28303331942b0fddda24bfc"} Dec 01 15:54:40 crc kubenswrapper[4739]: I1201 15:54:40.385861 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"50a7f5a5-c069-4e82-97e0-ecccdf443c15","Type":"ContainerStarted","Data":"b7953f17b42fd14759c490e6aa631975eb42ff6910ca8d330205568a9eb64953"} Dec 01 15:54:40 crc kubenswrapper[4739]: I1201 15:54:40.386351 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 01 15:54:40 crc kubenswrapper[4739]: I1201 15:54:40.424258 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.424238646 podStartE2EDuration="2.424238646s" podCreationTimestamp="2025-12-01 15:54:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:54:40.411581565 +0000 UTC m=+1182.237327699" watchObservedRunningTime="2025-12-01 15:54:40.424238646 +0000 UTC m=+1182.249984750" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.397796 4739 generic.go:334] "Generic (PLEG): container finished" podID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerID="ca3820b86fb7dfa5f667c711afb764f9439cd7ced3610355bed4b0e880bdcfd1" exitCode=137 Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.397918 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b4ee7d9-baee-491b-8663-461e89028fa2","Type":"ContainerDied","Data":"ca3820b86fb7dfa5f667c711afb764f9439cd7ced3610355bed4b0e880bdcfd1"} Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.399056 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4b4ee7d9-baee-491b-8663-461e89028fa2","Type":"ContainerDied","Data":"2ccfcab29dbecabd502499798681b9bebacc81d0081c27995d1d6373ed54e4e1"} Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.399072 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ccfcab29dbecabd502499798681b9bebacc81d0081c27995d1d6373ed54e4e1" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.437220 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.499708 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b4ee7d9-baee-491b-8663-461e89028fa2-run-httpd\") pod \"4b4ee7d9-baee-491b-8663-461e89028fa2\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.499774 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-combined-ca-bundle\") pod \"4b4ee7d9-baee-491b-8663-461e89028fa2\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.500594 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-sg-core-conf-yaml\") pod \"4b4ee7d9-baee-491b-8663-461e89028fa2\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.500668 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbmtn\" (UniqueName: \"kubernetes.io/projected/4b4ee7d9-baee-491b-8663-461e89028fa2-kube-api-access-jbmtn\") pod \"4b4ee7d9-baee-491b-8663-461e89028fa2\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.500699 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b4ee7d9-baee-491b-8663-461e89028fa2-log-httpd\") pod \"4b4ee7d9-baee-491b-8663-461e89028fa2\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.500768 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-config-data\") pod \"4b4ee7d9-baee-491b-8663-461e89028fa2\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.500803 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-scripts\") pod \"4b4ee7d9-baee-491b-8663-461e89028fa2\" (UID: \"4b4ee7d9-baee-491b-8663-461e89028fa2\") " Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.501060 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b4ee7d9-baee-491b-8663-461e89028fa2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4b4ee7d9-baee-491b-8663-461e89028fa2" (UID: "4b4ee7d9-baee-491b-8663-461e89028fa2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.501300 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b4ee7d9-baee-491b-8663-461e89028fa2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4b4ee7d9-baee-491b-8663-461e89028fa2" (UID: "4b4ee7d9-baee-491b-8663-461e89028fa2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.501879 4739 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b4ee7d9-baee-491b-8663-461e89028fa2-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.501912 4739 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4b4ee7d9-baee-491b-8663-461e89028fa2-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.506354 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-scripts" (OuterVolumeSpecName: "scripts") pod "4b4ee7d9-baee-491b-8663-461e89028fa2" (UID: "4b4ee7d9-baee-491b-8663-461e89028fa2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.506407 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b4ee7d9-baee-491b-8663-461e89028fa2-kube-api-access-jbmtn" (OuterVolumeSpecName: "kube-api-access-jbmtn") pod "4b4ee7d9-baee-491b-8663-461e89028fa2" (UID: "4b4ee7d9-baee-491b-8663-461e89028fa2"). InnerVolumeSpecName "kube-api-access-jbmtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.537140 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4b4ee7d9-baee-491b-8663-461e89028fa2" (UID: "4b4ee7d9-baee-491b-8663-461e89028fa2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.574593 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b4ee7d9-baee-491b-8663-461e89028fa2" (UID: "4b4ee7d9-baee-491b-8663-461e89028fa2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.603438 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.603468 4739 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.603476 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbmtn\" (UniqueName: \"kubernetes.io/projected/4b4ee7d9-baee-491b-8663-461e89028fa2-kube-api-access-jbmtn\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.603509 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.625686 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-config-data" (OuterVolumeSpecName: "config-data") pod "4b4ee7d9-baee-491b-8663-461e89028fa2" (UID: "4b4ee7d9-baee-491b-8663-461e89028fa2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:54:41 crc kubenswrapper[4739]: I1201 15:54:41.705225 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b4ee7d9-baee-491b-8663-461e89028fa2-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.409855 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.468898 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.497957 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.508801 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:54:42 crc kubenswrapper[4739]: E1201 15:54:42.509196 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="ceilometer-notification-agent" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.509219 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="ceilometer-notification-agent" Dec 01 15:54:42 crc kubenswrapper[4739]: E1201 15:54:42.509263 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="ceilometer-central-agent" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.509273 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="ceilometer-central-agent" Dec 01 15:54:42 crc kubenswrapper[4739]: E1201 15:54:42.509294 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="proxy-httpd" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.509303 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="proxy-httpd" Dec 01 15:54:42 crc kubenswrapper[4739]: E1201 15:54:42.509322 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="sg-core" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.509332 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="sg-core" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.510141 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="sg-core" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.510181 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="proxy-httpd" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.510199 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="ceilometer-notification-agent" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.510394 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" containerName="ceilometer-central-agent" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.512648 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.516171 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.516398 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.541551 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.619771 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t245v\" (UniqueName: \"kubernetes.io/projected/33904e6b-d018-4a30-a9c4-f57657f1795c-kube-api-access-t245v\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.619830 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.619850 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-scripts\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.620005 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-config-data\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.620087 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33904e6b-d018-4a30-a9c4-f57657f1795c-run-httpd\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.620184 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33904e6b-d018-4a30-a9c4-f57657f1795c-log-httpd\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.620279 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.722127 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33904e6b-d018-4a30-a9c4-f57657f1795c-log-httpd\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.722638 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.722874 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t245v\" (UniqueName: \"kubernetes.io/projected/33904e6b-d018-4a30-a9c4-f57657f1795c-kube-api-access-t245v\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.723086 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.722879 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33904e6b-d018-4a30-a9c4-f57657f1795c-log-httpd\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.723242 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-scripts\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.723636 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-config-data\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.723739 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33904e6b-d018-4a30-a9c4-f57657f1795c-run-httpd\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.724236 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33904e6b-d018-4a30-a9c4-f57657f1795c-run-httpd\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.728495 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.729769 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-scripts\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.730105 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-config-data\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.732794 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.744735 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t245v\" (UniqueName: \"kubernetes.io/projected/33904e6b-d018-4a30-a9c4-f57657f1795c-kube-api-access-t245v\") pod \"ceilometer-0\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " pod="openstack/ceilometer-0" Dec 01 15:54:42 crc kubenswrapper[4739]: I1201 15:54:42.835500 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:54:43 crc kubenswrapper[4739]: I1201 15:54:43.288861 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:54:43 crc kubenswrapper[4739]: W1201 15:54:43.293376 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33904e6b_d018_4a30_a9c4_f57657f1795c.slice/crio-995329ae8eb8a79e32d2310a95c4ff4383ef4d895dce629d586cedd7e8ab20a0 WatchSource:0}: Error finding container 995329ae8eb8a79e32d2310a95c4ff4383ef4d895dce629d586cedd7e8ab20a0: Status 404 returned error can't find the container with id 995329ae8eb8a79e32d2310a95c4ff4383ef4d895dce629d586cedd7e8ab20a0 Dec 01 15:54:43 crc kubenswrapper[4739]: I1201 15:54:43.420493 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33904e6b-d018-4a30-a9c4-f57657f1795c","Type":"ContainerStarted","Data":"995329ae8eb8a79e32d2310a95c4ff4383ef4d895dce629d586cedd7e8ab20a0"} Dec 01 15:54:44 crc kubenswrapper[4739]: I1201 15:54:44.435062 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33904e6b-d018-4a30-a9c4-f57657f1795c","Type":"ContainerStarted","Data":"774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df"} Dec 01 15:54:44 crc kubenswrapper[4739]: I1201 15:54:44.510289 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b4ee7d9-baee-491b-8663-461e89028fa2" path="/var/lib/kubelet/pods/4b4ee7d9-baee-491b-8663-461e89028fa2/volumes" Dec 01 15:54:45 crc kubenswrapper[4739]: I1201 15:54:45.446567 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33904e6b-d018-4a30-a9c4-f57657f1795c","Type":"ContainerStarted","Data":"73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560"} Dec 01 15:54:46 crc kubenswrapper[4739]: I1201 15:54:46.463073 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33904e6b-d018-4a30-a9c4-f57657f1795c","Type":"ContainerStarted","Data":"4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0"} Dec 01 15:54:48 crc kubenswrapper[4739]: I1201 15:54:48.505880 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33904e6b-d018-4a30-a9c4-f57657f1795c","Type":"ContainerStarted","Data":"9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0"} Dec 01 15:54:48 crc kubenswrapper[4739]: I1201 15:54:48.568583 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.030200274 podStartE2EDuration="6.568558255s" podCreationTimestamp="2025-12-01 15:54:42 +0000 UTC" firstStartedPulling="2025-12-01 15:54:43.29639687 +0000 UTC m=+1185.122142964" lastFinishedPulling="2025-12-01 15:54:47.834754851 +0000 UTC m=+1189.660500945" observedRunningTime="2025-12-01 15:54:48.544265283 +0000 UTC m=+1190.370011457" watchObservedRunningTime="2025-12-01 15:54:48.568558255 +0000 UTC m=+1190.394304369" Dec 01 15:54:48 crc kubenswrapper[4739]: I1201 15:54:48.923562 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.413146 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-clm6v"] Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.414869 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.418275 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.418647 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.435332 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-clm6v"] Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.465110 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-clm6v\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.465167 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxdck\" (UniqueName: \"kubernetes.io/projected/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-kube-api-access-pxdck\") pod \"nova-cell0-cell-mapping-clm6v\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.465473 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-scripts\") pod \"nova-cell0-cell-mapping-clm6v\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.465584 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-config-data\") pod \"nova-cell0-cell-mapping-clm6v\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.497796 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.567648 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-clm6v\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.567710 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxdck\" (UniqueName: \"kubernetes.io/projected/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-kube-api-access-pxdck\") pod \"nova-cell0-cell-mapping-clm6v\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.568927 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-scripts\") pod \"nova-cell0-cell-mapping-clm6v\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.569007 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-config-data\") pod \"nova-cell0-cell-mapping-clm6v\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.578236 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-scripts\") pod \"nova-cell0-cell-mapping-clm6v\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.579064 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-clm6v\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.587202 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-config-data\") pod \"nova-cell0-cell-mapping-clm6v\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.602318 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxdck\" (UniqueName: \"kubernetes.io/projected/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-kube-api-access-pxdck\") pod \"nova-cell0-cell-mapping-clm6v\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.604611 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.606328 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.610966 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.672357 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.679952 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6bv5\" (UniqueName: \"kubernetes.io/projected/b6afdc34-badd-4095-aa08-f2350c611f9d-kube-api-access-l6bv5\") pod \"nova-api-0\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " pod="openstack/nova-api-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.680033 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6afdc34-badd-4095-aa08-f2350c611f9d-logs\") pod \"nova-api-0\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " pod="openstack/nova-api-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.680064 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6afdc34-badd-4095-aa08-f2350c611f9d-config-data\") pod \"nova-api-0\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " pod="openstack/nova-api-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.680095 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6afdc34-badd-4095-aa08-f2350c611f9d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " pod="openstack/nova-api-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.723004 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.731940 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.743790 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.767203 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.770116 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.782942 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6afdc34-badd-4095-aa08-f2350c611f9d-config-data\") pod \"nova-api-0\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " pod="openstack/nova-api-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.782996 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6afdc34-badd-4095-aa08-f2350c611f9d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " pod="openstack/nova-api-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.783051 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c384778-e1c6-4072-8a54-5f5402aad753-logs\") pod \"nova-metadata-0\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " pod="openstack/nova-metadata-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.783068 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96l5g\" (UniqueName: \"kubernetes.io/projected/2c384778-e1c6-4072-8a54-5f5402aad753-kube-api-access-96l5g\") pod \"nova-metadata-0\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " pod="openstack/nova-metadata-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.783092 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c384778-e1c6-4072-8a54-5f5402aad753-config-data\") pod \"nova-metadata-0\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " pod="openstack/nova-metadata-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.783117 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c384778-e1c6-4072-8a54-5f5402aad753-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " pod="openstack/nova-metadata-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.783139 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6bv5\" (UniqueName: \"kubernetes.io/projected/b6afdc34-badd-4095-aa08-f2350c611f9d-kube-api-access-l6bv5\") pod \"nova-api-0\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " pod="openstack/nova-api-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.783198 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6afdc34-badd-4095-aa08-f2350c611f9d-logs\") pod \"nova-api-0\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " pod="openstack/nova-api-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.783595 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6afdc34-badd-4095-aa08-f2350c611f9d-logs\") pod \"nova-api-0\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " pod="openstack/nova-api-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.792706 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-x89dh"] Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.794284 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.796007 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6afdc34-badd-4095-aa08-f2350c611f9d-config-data\") pod \"nova-api-0\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " pod="openstack/nova-api-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.801075 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6afdc34-badd-4095-aa08-f2350c611f9d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " pod="openstack/nova-api-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.809945 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-x89dh"] Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.848956 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6bv5\" (UniqueName: \"kubernetes.io/projected/b6afdc34-badd-4095-aa08-f2350c611f9d-kube-api-access-l6bv5\") pod \"nova-api-0\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " pod="openstack/nova-api-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.884817 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-x89dh\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.885049 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwfp4\" (UniqueName: \"kubernetes.io/projected/eb53ff5e-3cee-4686-befd-d071117f4541-kube-api-access-dwfp4\") pod \"dnsmasq-dns-566b5b7845-x89dh\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.885075 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-dns-svc\") pod \"dnsmasq-dns-566b5b7845-x89dh\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.885137 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c384778-e1c6-4072-8a54-5f5402aad753-logs\") pod \"nova-metadata-0\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " pod="openstack/nova-metadata-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.885156 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96l5g\" (UniqueName: \"kubernetes.io/projected/2c384778-e1c6-4072-8a54-5f5402aad753-kube-api-access-96l5g\") pod \"nova-metadata-0\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " pod="openstack/nova-metadata-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.885179 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c384778-e1c6-4072-8a54-5f5402aad753-config-data\") pod \"nova-metadata-0\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " pod="openstack/nova-metadata-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.885201 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-config\") pod \"dnsmasq-dns-566b5b7845-x89dh\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.885218 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-x89dh\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.885239 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c384778-e1c6-4072-8a54-5f5402aad753-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " pod="openstack/nova-metadata-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.886734 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c384778-e1c6-4072-8a54-5f5402aad753-logs\") pod \"nova-metadata-0\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " pod="openstack/nova-metadata-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.895738 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c384778-e1c6-4072-8a54-5f5402aad753-config-data\") pod \"nova-metadata-0\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " pod="openstack/nova-metadata-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.899773 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c384778-e1c6-4072-8a54-5f5402aad753-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " pod="openstack/nova-metadata-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.905411 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.906731 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.910051 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.935880 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96l5g\" (UniqueName: \"kubernetes.io/projected/2c384778-e1c6-4072-8a54-5f5402aad753-kube-api-access-96l5g\") pod \"nova-metadata-0\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " pod="openstack/nova-metadata-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.940138 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.959579 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.960752 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.964295 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.988284 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwfp4\" (UniqueName: \"kubernetes.io/projected/eb53ff5e-3cee-4686-befd-d071117f4541-kube-api-access-dwfp4\") pod \"dnsmasq-dns-566b5b7845-x89dh\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.988341 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-dns-svc\") pod \"dnsmasq-dns-566b5b7845-x89dh\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.988372 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d9292e1-fbbf-46d2-85ad-4eda18061dbb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.988435 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc4ns\" (UniqueName: \"kubernetes.io/projected/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-kube-api-access-mc4ns\") pod \"nova-scheduler-0\" (UID: \"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf\") " pod="openstack/nova-scheduler-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.988490 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-config\") pod \"dnsmasq-dns-566b5b7845-x89dh\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.988521 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-x89dh\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.988539 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pklg\" (UniqueName: \"kubernetes.io/projected/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-kube-api-access-2pklg\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d9292e1-fbbf-46d2-85ad-4eda18061dbb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.988583 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf\") " pod="openstack/nova-scheduler-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.988601 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d9292e1-fbbf-46d2-85ad-4eda18061dbb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.988643 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-x89dh\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.988658 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-config-data\") pod \"nova-scheduler-0\" (UID: \"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf\") " pod="openstack/nova-scheduler-0" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.989608 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-ovsdbserver-nb\") pod \"dnsmasq-dns-566b5b7845-x89dh\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.989820 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-dns-svc\") pod \"dnsmasq-dns-566b5b7845-x89dh\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:49 crc kubenswrapper[4739]: I1201 15:54:49.990250 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-ovsdbserver-sb\") pod \"dnsmasq-dns-566b5b7845-x89dh\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.000837 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-config\") pod \"dnsmasq-dns-566b5b7845-x89dh\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.018391 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.024843 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwfp4\" (UniqueName: \"kubernetes.io/projected/eb53ff5e-3cee-4686-befd-d071117f4541-kube-api-access-dwfp4\") pod \"dnsmasq-dns-566b5b7845-x89dh\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.030469 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.092993 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.093374 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf\") " pod="openstack/nova-scheduler-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.093434 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d9292e1-fbbf-46d2-85ad-4eda18061dbb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.093481 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-config-data\") pod \"nova-scheduler-0\" (UID: \"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf\") " pod="openstack/nova-scheduler-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.093525 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d9292e1-fbbf-46d2-85ad-4eda18061dbb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.093572 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc4ns\" (UniqueName: \"kubernetes.io/projected/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-kube-api-access-mc4ns\") pod \"nova-scheduler-0\" (UID: \"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf\") " pod="openstack/nova-scheduler-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.093617 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pklg\" (UniqueName: \"kubernetes.io/projected/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-kube-api-access-2pklg\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d9292e1-fbbf-46d2-85ad-4eda18061dbb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.098643 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d9292e1-fbbf-46d2-85ad-4eda18061dbb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.098782 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf\") " pod="openstack/nova-scheduler-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.099541 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d9292e1-fbbf-46d2-85ad-4eda18061dbb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.105348 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-config-data\") pod \"nova-scheduler-0\" (UID: \"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf\") " pod="openstack/nova-scheduler-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.116578 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc4ns\" (UniqueName: \"kubernetes.io/projected/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-kube-api-access-mc4ns\") pod \"nova-scheduler-0\" (UID: \"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf\") " pod="openstack/nova-scheduler-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.120714 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pklg\" (UniqueName: \"kubernetes.io/projected/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-kube-api-access-2pklg\") pod \"nova-cell1-novncproxy-0\" (UID: \"7d9292e1-fbbf-46d2-85ad-4eda18061dbb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.262049 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.288080 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.294232 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.554758 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-clm6v"] Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.575639 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-t9wxt"] Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.576745 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.581067 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.581325 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.597612 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-t9wxt"] Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.612482 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-t9wxt\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.612528 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-scripts\") pod \"nova-cell1-conductor-db-sync-t9wxt\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.612552 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz2cx\" (UniqueName: \"kubernetes.io/projected/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-kube-api-access-gz2cx\") pod \"nova-cell1-conductor-db-sync-t9wxt\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.612669 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-config-data\") pod \"nova-cell1-conductor-db-sync-t9wxt\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.651749 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.666987 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:54:50 crc kubenswrapper[4739]: W1201 15:54:50.672658 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c384778_e1c6_4072_8a54_5f5402aad753.slice/crio-cf89affc72716bcda000a700e2114859e52fd6d177776b497e03acb7b17f11a6 WatchSource:0}: Error finding container cf89affc72716bcda000a700e2114859e52fd6d177776b497e03acb7b17f11a6: Status 404 returned error can't find the container with id cf89affc72716bcda000a700e2114859e52fd6d177776b497e03acb7b17f11a6 Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.713906 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-t9wxt\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.713950 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-scripts\") pod \"nova-cell1-conductor-db-sync-t9wxt\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.713968 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz2cx\" (UniqueName: \"kubernetes.io/projected/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-kube-api-access-gz2cx\") pod \"nova-cell1-conductor-db-sync-t9wxt\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.714041 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-config-data\") pod \"nova-cell1-conductor-db-sync-t9wxt\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.719598 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-scripts\") pod \"nova-cell1-conductor-db-sync-t9wxt\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.731734 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-config-data\") pod \"nova-cell1-conductor-db-sync-t9wxt\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.736458 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-t9wxt\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.737445 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz2cx\" (UniqueName: \"kubernetes.io/projected/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-kube-api-access-gz2cx\") pod \"nova-cell1-conductor-db-sync-t9wxt\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.800866 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.829961 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-x89dh"] Dec 01 15:54:50 crc kubenswrapper[4739]: W1201 15:54:50.830784 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb53ff5e_3cee_4686_befd_d071117f4541.slice/crio-09e5f232ae637ff01321a16712f4dc7ddd20fc1dbbc6d4b5f2b67e1dc40c826d WatchSource:0}: Error finding container 09e5f232ae637ff01321a16712f4dc7ddd20fc1dbbc6d4b5f2b67e1dc40c826d: Status 404 returned error can't find the container with id 09e5f232ae637ff01321a16712f4dc7ddd20fc1dbbc6d4b5f2b67e1dc40c826d Dec 01 15:54:50 crc kubenswrapper[4739]: I1201 15:54:50.955231 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 15:54:51 crc kubenswrapper[4739]: I1201 15:54:51.054484 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 15:54:51 crc kubenswrapper[4739]: W1201 15:54:51.062756 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d9292e1_fbbf_46d2_85ad_4eda18061dbb.slice/crio-1da04fcb50be5c944f5302f8c71eac4662e15ca1ed533e97f7baa5e620bb211b WatchSource:0}: Error finding container 1da04fcb50be5c944f5302f8c71eac4662e15ca1ed533e97f7baa5e620bb211b: Status 404 returned error can't find the container with id 1da04fcb50be5c944f5302f8c71eac4662e15ca1ed533e97f7baa5e620bb211b Dec 01 15:54:51 crc kubenswrapper[4739]: I1201 15:54:51.233985 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-t9wxt"] Dec 01 15:54:51 crc kubenswrapper[4739]: W1201 15:54:51.234068 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1af0156_b7e0_4257_a1ad_2ae110ae8d54.slice/crio-59cba5581f5042ada44d0ac5a51d766a8eb8a9a7a67da7ac1e2e4784c1205d78 WatchSource:0}: Error finding container 59cba5581f5042ada44d0ac5a51d766a8eb8a9a7a67da7ac1e2e4784c1205d78: Status 404 returned error can't find the container with id 59cba5581f5042ada44d0ac5a51d766a8eb8a9a7a67da7ac1e2e4784c1205d78 Dec 01 15:54:51 crc kubenswrapper[4739]: I1201 15:54:51.536520 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c384778-e1c6-4072-8a54-5f5402aad753","Type":"ContainerStarted","Data":"cf89affc72716bcda000a700e2114859e52fd6d177776b497e03acb7b17f11a6"} Dec 01 15:54:51 crc kubenswrapper[4739]: I1201 15:54:51.538577 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b6afdc34-badd-4095-aa08-f2350c611f9d","Type":"ContainerStarted","Data":"591d513ba70da4b468eaca79781d6f6799b99367a93331911c275a4e4c7e699c"} Dec 01 15:54:51 crc kubenswrapper[4739]: I1201 15:54:51.540810 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-clm6v" event={"ID":"e3b310c6-127c-45ba-be81-bb6cd95fcb7b","Type":"ContainerStarted","Data":"919e6ab567e8b3ac14962fa8ab6d6a986cd9e8f3d3b1db1c46dc5db76295131c"} Dec 01 15:54:51 crc kubenswrapper[4739]: I1201 15:54:51.540842 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-clm6v" event={"ID":"e3b310c6-127c-45ba-be81-bb6cd95fcb7b","Type":"ContainerStarted","Data":"f0b8b189b77b10455ef60707681b9e17920f6cdc532959b2c4155ecb9a98c22d"} Dec 01 15:54:51 crc kubenswrapper[4739]: I1201 15:54:51.542254 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf","Type":"ContainerStarted","Data":"e69781824609bd98645ab94c4295aca272bc4312c159ff634e1d46d63e803b1a"} Dec 01 15:54:51 crc kubenswrapper[4739]: I1201 15:54:51.544812 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7d9292e1-fbbf-46d2-85ad-4eda18061dbb","Type":"ContainerStarted","Data":"1da04fcb50be5c944f5302f8c71eac4662e15ca1ed533e97f7baa5e620bb211b"} Dec 01 15:54:51 crc kubenswrapper[4739]: I1201 15:54:51.548466 4739 generic.go:334] "Generic (PLEG): container finished" podID="eb53ff5e-3cee-4686-befd-d071117f4541" containerID="fdb7a80be083b941db94ed60ecef17fc6a9bad38033f2342ba4202e43525f688" exitCode=0 Dec 01 15:54:51 crc kubenswrapper[4739]: I1201 15:54:51.548542 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-x89dh" event={"ID":"eb53ff5e-3cee-4686-befd-d071117f4541","Type":"ContainerDied","Data":"fdb7a80be083b941db94ed60ecef17fc6a9bad38033f2342ba4202e43525f688"} Dec 01 15:54:51 crc kubenswrapper[4739]: I1201 15:54:51.548564 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-x89dh" event={"ID":"eb53ff5e-3cee-4686-befd-d071117f4541","Type":"ContainerStarted","Data":"09e5f232ae637ff01321a16712f4dc7ddd20fc1dbbc6d4b5f2b67e1dc40c826d"} Dec 01 15:54:51 crc kubenswrapper[4739]: I1201 15:54:51.549597 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-t9wxt" event={"ID":"a1af0156-b7e0-4257-a1ad-2ae110ae8d54","Type":"ContainerStarted","Data":"19165388877e6a8140fc0e43fbaa05e81a77f89bed4844ade58d6f86738dc28c"} Dec 01 15:54:51 crc kubenswrapper[4739]: I1201 15:54:51.549631 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-t9wxt" event={"ID":"a1af0156-b7e0-4257-a1ad-2ae110ae8d54","Type":"ContainerStarted","Data":"59cba5581f5042ada44d0ac5a51d766a8eb8a9a7a67da7ac1e2e4784c1205d78"} Dec 01 15:54:51 crc kubenswrapper[4739]: I1201 15:54:51.561465 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-clm6v" podStartSLOduration=2.561449307 podStartE2EDuration="2.561449307s" podCreationTimestamp="2025-12-01 15:54:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:54:51.558469895 +0000 UTC m=+1193.384215989" watchObservedRunningTime="2025-12-01 15:54:51.561449307 +0000 UTC m=+1193.387195401" Dec 01 15:54:51 crc kubenswrapper[4739]: I1201 15:54:51.607239 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-t9wxt" podStartSLOduration=1.607210345 podStartE2EDuration="1.607210345s" podCreationTimestamp="2025-12-01 15:54:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:54:51.593047916 +0000 UTC m=+1193.418794020" watchObservedRunningTime="2025-12-01 15:54:51.607210345 +0000 UTC m=+1193.432956439" Dec 01 15:54:52 crc kubenswrapper[4739]: I1201 15:54:52.562592 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-x89dh" event={"ID":"eb53ff5e-3cee-4686-befd-d071117f4541","Type":"ContainerStarted","Data":"08302d554bc32eb4521c43e320f19d3a45fbac9b4964ec745918cfee15b5965a"} Dec 01 15:54:52 crc kubenswrapper[4739]: I1201 15:54:52.563311 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:54:52 crc kubenswrapper[4739]: I1201 15:54:52.583287 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-566b5b7845-x89dh" podStartSLOduration=3.5832690400000002 podStartE2EDuration="3.58326904s" podCreationTimestamp="2025-12-01 15:54:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:54:52.582128055 +0000 UTC m=+1194.407874149" watchObservedRunningTime="2025-12-01 15:54:52.58326904 +0000 UTC m=+1194.409015134" Dec 01 15:54:53 crc kubenswrapper[4739]: I1201 15:54:53.403121 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:54:53 crc kubenswrapper[4739]: I1201 15:54:53.411718 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 15:54:54 crc kubenswrapper[4739]: I1201 15:54:54.582355 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c384778-e1c6-4072-8a54-5f5402aad753","Type":"ContainerStarted","Data":"cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91"} Dec 01 15:54:54 crc kubenswrapper[4739]: I1201 15:54:54.583528 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c384778-e1c6-4072-8a54-5f5402aad753","Type":"ContainerStarted","Data":"6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579"} Dec 01 15:54:54 crc kubenswrapper[4739]: I1201 15:54:54.582637 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2c384778-e1c6-4072-8a54-5f5402aad753" containerName="nova-metadata-metadata" containerID="cri-o://cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91" gracePeriod=30 Dec 01 15:54:54 crc kubenswrapper[4739]: I1201 15:54:54.582525 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2c384778-e1c6-4072-8a54-5f5402aad753" containerName="nova-metadata-log" containerID="cri-o://6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579" gracePeriod=30 Dec 01 15:54:54 crc kubenswrapper[4739]: I1201 15:54:54.584895 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b6afdc34-badd-4095-aa08-f2350c611f9d","Type":"ContainerStarted","Data":"d0b2941b270052328077bd733a15a6d7e32636d3f8b755640441e218a1bc7a1c"} Dec 01 15:54:54 crc kubenswrapper[4739]: I1201 15:54:54.584971 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b6afdc34-badd-4095-aa08-f2350c611f9d","Type":"ContainerStarted","Data":"2ab24ed3341c32b08df4ea109ef756de9b0ae9f562035ae1631724747467ce78"} Dec 01 15:54:54 crc kubenswrapper[4739]: I1201 15:54:54.587826 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf","Type":"ContainerStarted","Data":"7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d"} Dec 01 15:54:54 crc kubenswrapper[4739]: I1201 15:54:54.589966 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="7d9292e1-fbbf-46d2-85ad-4eda18061dbb" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://0b432b1274db4acbebc79ac56068c6a2a2fc8a58426387136bdb5c2b3a828ab2" gracePeriod=30 Dec 01 15:54:54 crc kubenswrapper[4739]: I1201 15:54:54.590110 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7d9292e1-fbbf-46d2-85ad-4eda18061dbb","Type":"ContainerStarted","Data":"0b432b1274db4acbebc79ac56068c6a2a2fc8a58426387136bdb5c2b3a828ab2"} Dec 01 15:54:54 crc kubenswrapper[4739]: I1201 15:54:54.641151 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.05689803 podStartE2EDuration="5.641125907s" podCreationTimestamp="2025-12-01 15:54:49 +0000 UTC" firstStartedPulling="2025-12-01 15:54:51.068254194 +0000 UTC m=+1192.894000288" lastFinishedPulling="2025-12-01 15:54:53.652482051 +0000 UTC m=+1195.478228165" observedRunningTime="2025-12-01 15:54:54.637876576 +0000 UTC m=+1196.463622670" watchObservedRunningTime="2025-12-01 15:54:54.641125907 +0000 UTC m=+1196.466872001" Dec 01 15:54:54 crc kubenswrapper[4739]: I1201 15:54:54.648409 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.6732967199999997 podStartE2EDuration="5.648382501s" podCreationTimestamp="2025-12-01 15:54:49 +0000 UTC" firstStartedPulling="2025-12-01 15:54:50.677605567 +0000 UTC m=+1192.503351661" lastFinishedPulling="2025-12-01 15:54:53.652691348 +0000 UTC m=+1195.478437442" observedRunningTime="2025-12-01 15:54:54.61635399 +0000 UTC m=+1196.442100084" watchObservedRunningTime="2025-12-01 15:54:54.648382501 +0000 UTC m=+1196.474128595" Dec 01 15:54:54 crc kubenswrapper[4739]: I1201 15:54:54.659243 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.9832163879999998 podStartE2EDuration="5.659227797s" podCreationTimestamp="2025-12-01 15:54:49 +0000 UTC" firstStartedPulling="2025-12-01 15:54:50.981498857 +0000 UTC m=+1192.807244951" lastFinishedPulling="2025-12-01 15:54:53.657510266 +0000 UTC m=+1195.483256360" observedRunningTime="2025-12-01 15:54:54.65543912 +0000 UTC m=+1196.481185204" watchObservedRunningTime="2025-12-01 15:54:54.659227797 +0000 UTC m=+1196.484973891" Dec 01 15:54:54 crc kubenswrapper[4739]: I1201 15:54:54.683411 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.724584859 podStartE2EDuration="5.683388816s" podCreationTimestamp="2025-12-01 15:54:49 +0000 UTC" firstStartedPulling="2025-12-01 15:54:50.695873972 +0000 UTC m=+1192.521620076" lastFinishedPulling="2025-12-01 15:54:53.654677939 +0000 UTC m=+1195.480424033" observedRunningTime="2025-12-01 15:54:54.677773572 +0000 UTC m=+1196.503519666" watchObservedRunningTime="2025-12-01 15:54:54.683388816 +0000 UTC m=+1196.509134910" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.093762 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.093817 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.207348 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.289136 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.295347 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.317190 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96l5g\" (UniqueName: \"kubernetes.io/projected/2c384778-e1c6-4072-8a54-5f5402aad753-kube-api-access-96l5g\") pod \"2c384778-e1c6-4072-8a54-5f5402aad753\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.317228 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c384778-e1c6-4072-8a54-5f5402aad753-config-data\") pod \"2c384778-e1c6-4072-8a54-5f5402aad753\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.317245 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c384778-e1c6-4072-8a54-5f5402aad753-combined-ca-bundle\") pod \"2c384778-e1c6-4072-8a54-5f5402aad753\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.317374 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c384778-e1c6-4072-8a54-5f5402aad753-logs\") pod \"2c384778-e1c6-4072-8a54-5f5402aad753\" (UID: \"2c384778-e1c6-4072-8a54-5f5402aad753\") " Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.318124 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c384778-e1c6-4072-8a54-5f5402aad753-logs" (OuterVolumeSpecName: "logs") pod "2c384778-e1c6-4072-8a54-5f5402aad753" (UID: "2c384778-e1c6-4072-8a54-5f5402aad753"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.323490 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c384778-e1c6-4072-8a54-5f5402aad753-kube-api-access-96l5g" (OuterVolumeSpecName: "kube-api-access-96l5g") pod "2c384778-e1c6-4072-8a54-5f5402aad753" (UID: "2c384778-e1c6-4072-8a54-5f5402aad753"). InnerVolumeSpecName "kube-api-access-96l5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.348144 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c384778-e1c6-4072-8a54-5f5402aad753-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c384778-e1c6-4072-8a54-5f5402aad753" (UID: "2c384778-e1c6-4072-8a54-5f5402aad753"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.348641 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c384778-e1c6-4072-8a54-5f5402aad753-config-data" (OuterVolumeSpecName: "config-data") pod "2c384778-e1c6-4072-8a54-5f5402aad753" (UID: "2c384778-e1c6-4072-8a54-5f5402aad753"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.420538 4739 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c384778-e1c6-4072-8a54-5f5402aad753-logs\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.420597 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96l5g\" (UniqueName: \"kubernetes.io/projected/2c384778-e1c6-4072-8a54-5f5402aad753-kube-api-access-96l5g\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.420624 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c384778-e1c6-4072-8a54-5f5402aad753-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.420652 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c384778-e1c6-4072-8a54-5f5402aad753-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.604105 4739 generic.go:334] "Generic (PLEG): container finished" podID="2c384778-e1c6-4072-8a54-5f5402aad753" containerID="cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91" exitCode=0 Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.604146 4739 generic.go:334] "Generic (PLEG): container finished" podID="2c384778-e1c6-4072-8a54-5f5402aad753" containerID="6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579" exitCode=143 Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.604164 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c384778-e1c6-4072-8a54-5f5402aad753","Type":"ContainerDied","Data":"cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91"} Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.604247 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c384778-e1c6-4072-8a54-5f5402aad753","Type":"ContainerDied","Data":"6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579"} Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.604268 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c384778-e1c6-4072-8a54-5f5402aad753","Type":"ContainerDied","Data":"cf89affc72716bcda000a700e2114859e52fd6d177776b497e03acb7b17f11a6"} Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.604290 4739 scope.go:117] "RemoveContainer" containerID="cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.604338 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.647387 4739 scope.go:117] "RemoveContainer" containerID="6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.669263 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.687231 4739 scope.go:117] "RemoveContainer" containerID="cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.689210 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:54:55 crc kubenswrapper[4739]: E1201 15:54:55.695753 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91\": container with ID starting with cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91 not found: ID does not exist" containerID="cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.695809 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91"} err="failed to get container status \"cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91\": rpc error: code = NotFound desc = could not find container \"cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91\": container with ID starting with cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91 not found: ID does not exist" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.695836 4739 scope.go:117] "RemoveContainer" containerID="6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579" Dec 01 15:54:55 crc kubenswrapper[4739]: E1201 15:54:55.698716 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579\": container with ID starting with 6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579 not found: ID does not exist" containerID="6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.698748 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579"} err="failed to get container status \"6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579\": rpc error: code = NotFound desc = could not find container \"6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579\": container with ID starting with 6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579 not found: ID does not exist" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.698766 4739 scope.go:117] "RemoveContainer" containerID="cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.698846 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:54:55 crc kubenswrapper[4739]: E1201 15:54:55.699274 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c384778-e1c6-4072-8a54-5f5402aad753" containerName="nova-metadata-log" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.699290 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c384778-e1c6-4072-8a54-5f5402aad753" containerName="nova-metadata-log" Dec 01 15:54:55 crc kubenswrapper[4739]: E1201 15:54:55.699313 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c384778-e1c6-4072-8a54-5f5402aad753" containerName="nova-metadata-metadata" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.699321 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c384778-e1c6-4072-8a54-5f5402aad753" containerName="nova-metadata-metadata" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.699532 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c384778-e1c6-4072-8a54-5f5402aad753" containerName="nova-metadata-log" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.699550 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c384778-e1c6-4072-8a54-5f5402aad753" containerName="nova-metadata-metadata" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.700607 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.701527 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91"} err="failed to get container status \"cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91\": rpc error: code = NotFound desc = could not find container \"cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91\": container with ID starting with cf777b4543cf2e303bcf1ec6fa04f2a9ff63f21332b735df7f0dbbc962ab1a91 not found: ID does not exist" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.701552 4739 scope.go:117] "RemoveContainer" containerID="6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.711952 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.712176 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.712379 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579"} err="failed to get container status \"6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579\": rpc error: code = NotFound desc = could not find container \"6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579\": container with ID starting with 6786a0848121a6ea1bc69ccbde31eee588da9570d268636d8b4128faf4136579 not found: ID does not exist" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.719575 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.832890 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.833177 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-config-data\") pod \"nova-metadata-0\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.833335 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rswsp\" (UniqueName: \"kubernetes.io/projected/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-kube-api-access-rswsp\") pod \"nova-metadata-0\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.833558 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.833687 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-logs\") pod \"nova-metadata-0\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.935737 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.935806 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-logs\") pod \"nova-metadata-0\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.935856 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.935889 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-config-data\") pod \"nova-metadata-0\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.935925 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rswsp\" (UniqueName: \"kubernetes.io/projected/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-kube-api-access-rswsp\") pod \"nova-metadata-0\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.936705 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-logs\") pod \"nova-metadata-0\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.941281 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.942288 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.949346 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-config-data\") pod \"nova-metadata-0\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " pod="openstack/nova-metadata-0" Dec 01 15:54:55 crc kubenswrapper[4739]: I1201 15:54:55.961773 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rswsp\" (UniqueName: \"kubernetes.io/projected/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-kube-api-access-rswsp\") pod \"nova-metadata-0\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " pod="openstack/nova-metadata-0" Dec 01 15:54:56 crc kubenswrapper[4739]: I1201 15:54:56.046665 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 15:54:56 crc kubenswrapper[4739]: I1201 15:54:56.492051 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c384778-e1c6-4072-8a54-5f5402aad753" path="/var/lib/kubelet/pods/2c384778-e1c6-4072-8a54-5f5402aad753/volumes" Dec 01 15:54:56 crc kubenswrapper[4739]: W1201 15:54:56.534064 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6d6c517_58e9_42c9_a9bf_8f9fd2514574.slice/crio-790538f2bf7ad4b1ca6c22efda9e79d4ed7b1de22fc2c1bd8e116a54f72ca95b WatchSource:0}: Error finding container 790538f2bf7ad4b1ca6c22efda9e79d4ed7b1de22fc2c1bd8e116a54f72ca95b: Status 404 returned error can't find the container with id 790538f2bf7ad4b1ca6c22efda9e79d4ed7b1de22fc2c1bd8e116a54f72ca95b Dec 01 15:54:56 crc kubenswrapper[4739]: I1201 15:54:56.543653 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:54:56 crc kubenswrapper[4739]: I1201 15:54:56.617753 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c6d6c517-58e9-42c9-a9bf-8f9fd2514574","Type":"ContainerStarted","Data":"790538f2bf7ad4b1ca6c22efda9e79d4ed7b1de22fc2c1bd8e116a54f72ca95b"} Dec 01 15:54:57 crc kubenswrapper[4739]: I1201 15:54:57.632723 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c6d6c517-58e9-42c9-a9bf-8f9fd2514574","Type":"ContainerStarted","Data":"3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f"} Dec 01 15:54:57 crc kubenswrapper[4739]: I1201 15:54:57.633277 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c6d6c517-58e9-42c9-a9bf-8f9fd2514574","Type":"ContainerStarted","Data":"e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb"} Dec 01 15:54:57 crc kubenswrapper[4739]: I1201 15:54:57.654935 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.654917906 podStartE2EDuration="2.654917906s" podCreationTimestamp="2025-12-01 15:54:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:54:57.65406563 +0000 UTC m=+1199.479811724" watchObservedRunningTime="2025-12-01 15:54:57.654917906 +0000 UTC m=+1199.480664020" Dec 01 15:54:57 crc kubenswrapper[4739]: E1201 15:54:57.995018 4739 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1af0156_b7e0_4257_a1ad_2ae110ae8d54.slice/crio-19165388877e6a8140fc0e43fbaa05e81a77f89bed4844ade58d6f86738dc28c.scope\": RecentStats: unable to find data in memory cache]" Dec 01 15:54:58 crc kubenswrapper[4739]: I1201 15:54:58.645611 4739 generic.go:334] "Generic (PLEG): container finished" podID="a1af0156-b7e0-4257-a1ad-2ae110ae8d54" containerID="19165388877e6a8140fc0e43fbaa05e81a77f89bed4844ade58d6f86738dc28c" exitCode=0 Dec 01 15:54:58 crc kubenswrapper[4739]: I1201 15:54:58.645713 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-t9wxt" event={"ID":"a1af0156-b7e0-4257-a1ad-2ae110ae8d54","Type":"ContainerDied","Data":"19165388877e6a8140fc0e43fbaa05e81a77f89bed4844ade58d6f86738dc28c"} Dec 01 15:54:58 crc kubenswrapper[4739]: I1201 15:54:58.648149 4739 generic.go:334] "Generic (PLEG): container finished" podID="e3b310c6-127c-45ba-be81-bb6cd95fcb7b" containerID="919e6ab567e8b3ac14962fa8ab6d6a986cd9e8f3d3b1db1c46dc5db76295131c" exitCode=0 Dec 01 15:54:58 crc kubenswrapper[4739]: I1201 15:54:58.648226 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-clm6v" event={"ID":"e3b310c6-127c-45ba-be81-bb6cd95fcb7b","Type":"ContainerDied","Data":"919e6ab567e8b3ac14962fa8ab6d6a986cd9e8f3d3b1db1c46dc5db76295131c"} Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.020077 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.020393 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.190145 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.201600 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.263626 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.291610 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.329519 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gz2cx\" (UniqueName: \"kubernetes.io/projected/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-kube-api-access-gz2cx\") pod \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.329633 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-combined-ca-bundle\") pod \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.329670 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-combined-ca-bundle\") pod \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.329695 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-config-data\") pod \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.329750 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-scripts\") pod \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\" (UID: \"a1af0156-b7e0-4257-a1ad-2ae110ae8d54\") " Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.329828 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-scripts\") pod \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.329887 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-config-data\") pod \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.329947 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxdck\" (UniqueName: \"kubernetes.io/projected/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-kube-api-access-pxdck\") pod \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\" (UID: \"e3b310c6-127c-45ba-be81-bb6cd95fcb7b\") " Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.331058 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.346638 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-scripts" (OuterVolumeSpecName: "scripts") pod "a1af0156-b7e0-4257-a1ad-2ae110ae8d54" (UID: "a1af0156-b7e0-4257-a1ad-2ae110ae8d54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.347259 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-kube-api-access-gz2cx" (OuterVolumeSpecName: "kube-api-access-gz2cx") pod "a1af0156-b7e0-4257-a1ad-2ae110ae8d54" (UID: "a1af0156-b7e0-4257-a1ad-2ae110ae8d54"). InnerVolumeSpecName "kube-api-access-gz2cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.347313 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-scripts" (OuterVolumeSpecName: "scripts") pod "e3b310c6-127c-45ba-be81-bb6cd95fcb7b" (UID: "e3b310c6-127c-45ba-be81-bb6cd95fcb7b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.347302 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-kube-api-access-pxdck" (OuterVolumeSpecName: "kube-api-access-pxdck") pod "e3b310c6-127c-45ba-be81-bb6cd95fcb7b" (UID: "e3b310c6-127c-45ba-be81-bb6cd95fcb7b"). InnerVolumeSpecName "kube-api-access-pxdck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.348550 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-rrz9k"] Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.348757 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" podUID="ee1eaf87-373f-4138-9625-2be87b6aaa7f" containerName="dnsmasq-dns" containerID="cri-o://16e6b948b38abbcd45d6da8fdbb50b5b859f6584246db6e9c48398ba169b07e4" gracePeriod=10 Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.371677 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-config-data" (OuterVolumeSpecName: "config-data") pod "e3b310c6-127c-45ba-be81-bb6cd95fcb7b" (UID: "e3b310c6-127c-45ba-be81-bb6cd95fcb7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.390545 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3b310c6-127c-45ba-be81-bb6cd95fcb7b" (UID: "e3b310c6-127c-45ba-be81-bb6cd95fcb7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.400903 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-config-data" (OuterVolumeSpecName: "config-data") pod "a1af0156-b7e0-4257-a1ad-2ae110ae8d54" (UID: "a1af0156-b7e0-4257-a1ad-2ae110ae8d54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.403810 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1af0156-b7e0-4257-a1ad-2ae110ae8d54" (UID: "a1af0156-b7e0-4257-a1ad-2ae110ae8d54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.431467 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.431493 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.431502 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.431510 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.431519 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.431528 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.431543 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxdck\" (UniqueName: \"kubernetes.io/projected/e3b310c6-127c-45ba-be81-bb6cd95fcb7b-kube-api-access-pxdck\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.431552 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gz2cx\" (UniqueName: \"kubernetes.io/projected/a1af0156-b7e0-4257-a1ad-2ae110ae8d54-kube-api-access-gz2cx\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.722947 4739 generic.go:334] "Generic (PLEG): container finished" podID="ee1eaf87-373f-4138-9625-2be87b6aaa7f" containerID="16e6b948b38abbcd45d6da8fdbb50b5b859f6584246db6e9c48398ba169b07e4" exitCode=0 Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.723190 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" event={"ID":"ee1eaf87-373f-4138-9625-2be87b6aaa7f","Type":"ContainerDied","Data":"16e6b948b38abbcd45d6da8fdbb50b5b859f6584246db6e9c48398ba169b07e4"} Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.724634 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-t9wxt" event={"ID":"a1af0156-b7e0-4257-a1ad-2ae110ae8d54","Type":"ContainerDied","Data":"59cba5581f5042ada44d0ac5a51d766a8eb8a9a7a67da7ac1e2e4784c1205d78"} Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.724655 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59cba5581f5042ada44d0ac5a51d766a8eb8a9a7a67da7ac1e2e4784c1205d78" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.724751 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-t9wxt" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.736021 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-clm6v" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.736982 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-clm6v" event={"ID":"e3b310c6-127c-45ba-be81-bb6cd95fcb7b","Type":"ContainerDied","Data":"f0b8b189b77b10455ef60707681b9e17920f6cdc532959b2c4155ecb9a98c22d"} Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.737010 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0b8b189b77b10455ef60707681b9e17920f6cdc532959b2c4155ecb9a98c22d" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.754067 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 15:55:00 crc kubenswrapper[4739]: E1201 15:55:00.754444 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b310c6-127c-45ba-be81-bb6cd95fcb7b" containerName="nova-manage" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.754457 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b310c6-127c-45ba-be81-bb6cd95fcb7b" containerName="nova-manage" Dec 01 15:55:00 crc kubenswrapper[4739]: E1201 15:55:00.754471 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1af0156-b7e0-4257-a1ad-2ae110ae8d54" containerName="nova-cell1-conductor-db-sync" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.754477 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1af0156-b7e0-4257-a1ad-2ae110ae8d54" containerName="nova-cell1-conductor-db-sync" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.754678 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3b310c6-127c-45ba-be81-bb6cd95fcb7b" containerName="nova-manage" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.754695 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1af0156-b7e0-4257-a1ad-2ae110ae8d54" containerName="nova-cell1-conductor-db-sync" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.755267 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.758864 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.770310 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.796483 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.840732 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/407b0796-c7a3-4ae4-903d-c8173b6cc378-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"407b0796-c7a3-4ae4-903d-c8173b6cc378\") " pod="openstack/nova-cell1-conductor-0" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.840778 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/407b0796-c7a3-4ae4-903d-c8173b6cc378-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"407b0796-c7a3-4ae4-903d-c8173b6cc378\") " pod="openstack/nova-cell1-conductor-0" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.840837 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9ds2\" (UniqueName: \"kubernetes.io/projected/407b0796-c7a3-4ae4-903d-c8173b6cc378-kube-api-access-b9ds2\") pod \"nova-cell1-conductor-0\" (UID: \"407b0796-c7a3-4ae4-903d-c8173b6cc378\") " pod="openstack/nova-cell1-conductor-0" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.852810 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.873327 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.873601 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b6afdc34-badd-4095-aa08-f2350c611f9d" containerName="nova-api-log" containerID="cri-o://2ab24ed3341c32b08df4ea109ef756de9b0ae9f562035ae1631724747467ce78" gracePeriod=30 Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.873642 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b6afdc34-badd-4095-aa08-f2350c611f9d" containerName="nova-api-api" containerID="cri-o://d0b2941b270052328077bd733a15a6d7e32636d3f8b755640441e218a1bc7a1c" gracePeriod=30 Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.881071 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b6afdc34-badd-4095-aa08-f2350c611f9d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.167:8774/\": EOF" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.881273 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b6afdc34-badd-4095-aa08-f2350c611f9d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.167:8774/\": EOF" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.893271 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.901413 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.901676 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c6d6c517-58e9-42c9-a9bf-8f9fd2514574" containerName="nova-metadata-log" containerID="cri-o://e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb" gracePeriod=30 Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.902076 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c6d6c517-58e9-42c9-a9bf-8f9fd2514574" containerName="nova-metadata-metadata" containerID="cri-o://3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f" gracePeriod=30 Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.941560 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-ovsdbserver-sb\") pod \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.941714 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4db64\" (UniqueName: \"kubernetes.io/projected/ee1eaf87-373f-4138-9625-2be87b6aaa7f-kube-api-access-4db64\") pod \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.941768 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-config\") pod \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.941824 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-dns-svc\") pod \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.941860 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-ovsdbserver-nb\") pod \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\" (UID: \"ee1eaf87-373f-4138-9625-2be87b6aaa7f\") " Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.942076 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/407b0796-c7a3-4ae4-903d-c8173b6cc378-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"407b0796-c7a3-4ae4-903d-c8173b6cc378\") " pod="openstack/nova-cell1-conductor-0" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.942095 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/407b0796-c7a3-4ae4-903d-c8173b6cc378-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"407b0796-c7a3-4ae4-903d-c8173b6cc378\") " pod="openstack/nova-cell1-conductor-0" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.942126 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9ds2\" (UniqueName: \"kubernetes.io/projected/407b0796-c7a3-4ae4-903d-c8173b6cc378-kube-api-access-b9ds2\") pod \"nova-cell1-conductor-0\" (UID: \"407b0796-c7a3-4ae4-903d-c8173b6cc378\") " pod="openstack/nova-cell1-conductor-0" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.948083 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/407b0796-c7a3-4ae4-903d-c8173b6cc378-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"407b0796-c7a3-4ae4-903d-c8173b6cc378\") " pod="openstack/nova-cell1-conductor-0" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.948598 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee1eaf87-373f-4138-9625-2be87b6aaa7f-kube-api-access-4db64" (OuterVolumeSpecName: "kube-api-access-4db64") pod "ee1eaf87-373f-4138-9625-2be87b6aaa7f" (UID: "ee1eaf87-373f-4138-9625-2be87b6aaa7f"). InnerVolumeSpecName "kube-api-access-4db64". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.952516 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/407b0796-c7a3-4ae4-903d-c8173b6cc378-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"407b0796-c7a3-4ae4-903d-c8173b6cc378\") " pod="openstack/nova-cell1-conductor-0" Dec 01 15:55:00 crc kubenswrapper[4739]: I1201 15:55:00.966553 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9ds2\" (UniqueName: \"kubernetes.io/projected/407b0796-c7a3-4ae4-903d-c8173b6cc378-kube-api-access-b9ds2\") pod \"nova-cell1-conductor-0\" (UID: \"407b0796-c7a3-4ae4-903d-c8173b6cc378\") " pod="openstack/nova-cell1-conductor-0" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.014163 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-config" (OuterVolumeSpecName: "config") pod "ee1eaf87-373f-4138-9625-2be87b6aaa7f" (UID: "ee1eaf87-373f-4138-9625-2be87b6aaa7f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.014323 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ee1eaf87-373f-4138-9625-2be87b6aaa7f" (UID: "ee1eaf87-373f-4138-9625-2be87b6aaa7f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.020032 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ee1eaf87-373f-4138-9625-2be87b6aaa7f" (UID: "ee1eaf87-373f-4138-9625-2be87b6aaa7f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.027281 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ee1eaf87-373f-4138-9625-2be87b6aaa7f" (UID: "ee1eaf87-373f-4138-9625-2be87b6aaa7f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.044344 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.044380 4739 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.044468 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.044481 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee1eaf87-373f-4138-9625-2be87b6aaa7f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.044492 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4db64\" (UniqueName: \"kubernetes.io/projected/ee1eaf87-373f-4138-9625-2be87b6aaa7f-kube-api-access-4db64\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.047533 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.047710 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.089919 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.508716 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.556081 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rswsp\" (UniqueName: \"kubernetes.io/projected/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-kube-api-access-rswsp\") pod \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.556146 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-nova-metadata-tls-certs\") pod \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.556189 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-config-data\") pod \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.556211 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-combined-ca-bundle\") pod \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.556281 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-logs\") pod \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\" (UID: \"c6d6c517-58e9-42c9-a9bf-8f9fd2514574\") " Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.562705 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-logs" (OuterVolumeSpecName: "logs") pod "c6d6c517-58e9-42c9-a9bf-8f9fd2514574" (UID: "c6d6c517-58e9-42c9-a9bf-8f9fd2514574"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.580678 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-kube-api-access-rswsp" (OuterVolumeSpecName: "kube-api-access-rswsp") pod "c6d6c517-58e9-42c9-a9bf-8f9fd2514574" (UID: "c6d6c517-58e9-42c9-a9bf-8f9fd2514574"). InnerVolumeSpecName "kube-api-access-rswsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.635641 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-config-data" (OuterVolumeSpecName: "config-data") pod "c6d6c517-58e9-42c9-a9bf-8f9fd2514574" (UID: "c6d6c517-58e9-42c9-a9bf-8f9fd2514574"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.658468 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rswsp\" (UniqueName: \"kubernetes.io/projected/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-kube-api-access-rswsp\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.658494 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.658504 4739 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-logs\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.668586 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6d6c517-58e9-42c9-a9bf-8f9fd2514574" (UID: "c6d6c517-58e9-42c9-a9bf-8f9fd2514574"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.674588 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c6d6c517-58e9-42c9-a9bf-8f9fd2514574" (UID: "c6d6c517-58e9-42c9-a9bf-8f9fd2514574"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.716548 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.761188 4739 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.761251 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6d6c517-58e9-42c9-a9bf-8f9fd2514574-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.761788 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"407b0796-c7a3-4ae4-903d-c8173b6cc378","Type":"ContainerStarted","Data":"6d946d6b6d274956286b9ba0dea5e078425210b34fd1a52da304985923042d78"} Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.765998 4739 generic.go:334] "Generic (PLEG): container finished" podID="c6d6c517-58e9-42c9-a9bf-8f9fd2514574" containerID="3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f" exitCode=0 Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.766028 4739 generic.go:334] "Generic (PLEG): container finished" podID="c6d6c517-58e9-42c9-a9bf-8f9fd2514574" containerID="e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb" exitCode=143 Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.766075 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c6d6c517-58e9-42c9-a9bf-8f9fd2514574","Type":"ContainerDied","Data":"3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f"} Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.766101 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c6d6c517-58e9-42c9-a9bf-8f9fd2514574","Type":"ContainerDied","Data":"e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb"} Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.766112 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c6d6c517-58e9-42c9-a9bf-8f9fd2514574","Type":"ContainerDied","Data":"790538f2bf7ad4b1ca6c22efda9e79d4ed7b1de22fc2c1bd8e116a54f72ca95b"} Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.766127 4739 scope.go:117] "RemoveContainer" containerID="3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.766246 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.775167 4739 generic.go:334] "Generic (PLEG): container finished" podID="b6afdc34-badd-4095-aa08-f2350c611f9d" containerID="2ab24ed3341c32b08df4ea109ef756de9b0ae9f562035ae1631724747467ce78" exitCode=143 Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.775261 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b6afdc34-badd-4095-aa08-f2350c611f9d","Type":"ContainerDied","Data":"2ab24ed3341c32b08df4ea109ef756de9b0ae9f562035ae1631724747467ce78"} Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.798212 4739 scope.go:117] "RemoveContainer" containerID="e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.799453 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" event={"ID":"ee1eaf87-373f-4138-9625-2be87b6aaa7f","Type":"ContainerDied","Data":"79c61688269609b18bda7f09a8b479ccaf967fbbe82dc8c2e70ec0c82c7184bb"} Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.799475 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d97fcdd8f-rrz9k" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.809124 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.820389 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.828394 4739 scope.go:117] "RemoveContainer" containerID="3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f" Dec 01 15:55:01 crc kubenswrapper[4739]: E1201 15:55:01.828842 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f\": container with ID starting with 3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f not found: ID does not exist" containerID="3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.828874 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f"} err="failed to get container status \"3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f\": rpc error: code = NotFound desc = could not find container \"3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f\": container with ID starting with 3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f not found: ID does not exist" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.828896 4739 scope.go:117] "RemoveContainer" containerID="e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb" Dec 01 15:55:01 crc kubenswrapper[4739]: E1201 15:55:01.832575 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb\": container with ID starting with e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb not found: ID does not exist" containerID="e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.832621 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb"} err="failed to get container status \"e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb\": rpc error: code = NotFound desc = could not find container \"e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb\": container with ID starting with e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb not found: ID does not exist" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.832647 4739 scope.go:117] "RemoveContainer" containerID="3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.836537 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.836601 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f"} err="failed to get container status \"3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f\": rpc error: code = NotFound desc = could not find container \"3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f\": container with ID starting with 3328df6d2ce80c312192c2d704134e13b363c28a458cb2271f2675109a4b0b0f not found: ID does not exist" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.836641 4739 scope.go:117] "RemoveContainer" containerID="e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb" Dec 01 15:55:01 crc kubenswrapper[4739]: E1201 15:55:01.836891 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee1eaf87-373f-4138-9625-2be87b6aaa7f" containerName="dnsmasq-dns" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.836907 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee1eaf87-373f-4138-9625-2be87b6aaa7f" containerName="dnsmasq-dns" Dec 01 15:55:01 crc kubenswrapper[4739]: E1201 15:55:01.836923 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6d6c517-58e9-42c9-a9bf-8f9fd2514574" containerName="nova-metadata-log" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.836929 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6d6c517-58e9-42c9-a9bf-8f9fd2514574" containerName="nova-metadata-log" Dec 01 15:55:01 crc kubenswrapper[4739]: E1201 15:55:01.836949 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6d6c517-58e9-42c9-a9bf-8f9fd2514574" containerName="nova-metadata-metadata" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.836955 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6d6c517-58e9-42c9-a9bf-8f9fd2514574" containerName="nova-metadata-metadata" Dec 01 15:55:01 crc kubenswrapper[4739]: E1201 15:55:01.836975 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee1eaf87-373f-4138-9625-2be87b6aaa7f" containerName="init" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.836981 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee1eaf87-373f-4138-9625-2be87b6aaa7f" containerName="init" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.837727 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee1eaf87-373f-4138-9625-2be87b6aaa7f" containerName="dnsmasq-dns" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.838334 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6d6c517-58e9-42c9-a9bf-8f9fd2514574" containerName="nova-metadata-log" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.838389 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6d6c517-58e9-42c9-a9bf-8f9fd2514574" containerName="nova-metadata-metadata" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.839575 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb"} err="failed to get container status \"e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb\": rpc error: code = NotFound desc = could not find container \"e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb\": container with ID starting with e975fd8080540b1ddccc8c8571058c50350350c1311ced6d53d90ffbc414a1fb not found: ID does not exist" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.839601 4739 scope.go:117] "RemoveContainer" containerID="16e6b948b38abbcd45d6da8fdbb50b5b859f6584246db6e9c48398ba169b07e4" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.846002 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.848942 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.856444 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.856683 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.872603 4739 scope.go:117] "RemoveContainer" containerID="397e83467e0d2793e2cdb6411b8ef238990c4bb95f28f6f23ca00cb4953c3457" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.872781 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-rrz9k"] Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.900670 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d97fcdd8f-rrz9k"] Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.963909 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " pod="openstack/nova-metadata-0" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.964095 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-config-data\") pod \"nova-metadata-0\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " pod="openstack/nova-metadata-0" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.964131 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " pod="openstack/nova-metadata-0" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.964149 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5dfq\" (UniqueName: \"kubernetes.io/projected/b30c462a-2b55-4187-931e-12e78990741a-kube-api-access-z5dfq\") pod \"nova-metadata-0\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " pod="openstack/nova-metadata-0" Dec 01 15:55:01 crc kubenswrapper[4739]: I1201 15:55:01.964179 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b30c462a-2b55-4187-931e-12e78990741a-logs\") pod \"nova-metadata-0\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " pod="openstack/nova-metadata-0" Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.066772 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " pod="openstack/nova-metadata-0" Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.069107 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-config-data\") pod \"nova-metadata-0\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " pod="openstack/nova-metadata-0" Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.069296 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " pod="openstack/nova-metadata-0" Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.069495 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5dfq\" (UniqueName: \"kubernetes.io/projected/b30c462a-2b55-4187-931e-12e78990741a-kube-api-access-z5dfq\") pod \"nova-metadata-0\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " pod="openstack/nova-metadata-0" Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.069692 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b30c462a-2b55-4187-931e-12e78990741a-logs\") pod \"nova-metadata-0\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " pod="openstack/nova-metadata-0" Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.071183 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " pod="openstack/nova-metadata-0" Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.071368 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b30c462a-2b55-4187-931e-12e78990741a-logs\") pod \"nova-metadata-0\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " pod="openstack/nova-metadata-0" Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.075666 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " pod="openstack/nova-metadata-0" Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.081894 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-config-data\") pod \"nova-metadata-0\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " pod="openstack/nova-metadata-0" Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.091091 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5dfq\" (UniqueName: \"kubernetes.io/projected/b30c462a-2b55-4187-931e-12e78990741a-kube-api-access-z5dfq\") pod \"nova-metadata-0\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " pod="openstack/nova-metadata-0" Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.179754 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.491566 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6d6c517-58e9-42c9-a9bf-8f9fd2514574" path="/var/lib/kubelet/pods/c6d6c517-58e9-42c9-a9bf-8f9fd2514574/volumes" Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.493146 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee1eaf87-373f-4138-9625-2be87b6aaa7f" path="/var/lib/kubelet/pods/ee1eaf87-373f-4138-9625-2be87b6aaa7f/volumes" Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.649821 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.813215 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"407b0796-c7a3-4ae4-903d-c8173b6cc378","Type":"ContainerStarted","Data":"7f86dc09ae66cecfae145b3f1ab045bd5024f39581d2da2cbe44058205c162fe"} Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.813311 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.814647 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b30c462a-2b55-4187-931e-12e78990741a","Type":"ContainerStarted","Data":"e50d0eb90a7b608f5c60145475ec10f7fe624510c0a3eaf5dd92d0796548bd06"} Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.815968 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1e139bd2-3b82-4d07-b96d-6d5b5ca947bf" containerName="nova-scheduler-scheduler" containerID="cri-o://7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d" gracePeriod=30 Dec 01 15:55:02 crc kubenswrapper[4739]: I1201 15:55:02.837702 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.8376788729999998 podStartE2EDuration="2.837678873s" podCreationTimestamp="2025-12-01 15:55:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:55:02.833355989 +0000 UTC m=+1204.659102083" watchObservedRunningTime="2025-12-01 15:55:02.837678873 +0000 UTC m=+1204.663424967" Dec 01 15:55:03 crc kubenswrapper[4739]: I1201 15:55:03.826265 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b30c462a-2b55-4187-931e-12e78990741a","Type":"ContainerStarted","Data":"6f0e116dd42b56eca373c2ab2adb6c497716ffc059119fd760c3616cb644f343"} Dec 01 15:55:03 crc kubenswrapper[4739]: I1201 15:55:03.826632 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b30c462a-2b55-4187-931e-12e78990741a","Type":"ContainerStarted","Data":"7c67238b1af7b9e356f869c4c76a43339bb2f33a8d2452ecd5d77211e6a33247"} Dec 01 15:55:05 crc kubenswrapper[4739]: E1201 15:55:05.290280 4739 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d is running failed: container process not found" containerID="7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 15:55:05 crc kubenswrapper[4739]: E1201 15:55:05.291089 4739 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d is running failed: container process not found" containerID="7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 15:55:05 crc kubenswrapper[4739]: E1201 15:55:05.291451 4739 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d is running failed: container process not found" containerID="7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 15:55:05 crc kubenswrapper[4739]: E1201 15:55:05.291520 4739 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1e139bd2-3b82-4d07-b96d-6d5b5ca947bf" containerName="nova-scheduler-scheduler" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.443612 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.470545 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.4705169940000005 podStartE2EDuration="4.470516994s" podCreationTimestamp="2025-12-01 15:55:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:55:03.853304184 +0000 UTC m=+1205.679050278" watchObservedRunningTime="2025-12-01 15:55:05.470516994 +0000 UTC m=+1207.296263108" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.534596 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc4ns\" (UniqueName: \"kubernetes.io/projected/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-kube-api-access-mc4ns\") pod \"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf\" (UID: \"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf\") " Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.534803 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-combined-ca-bundle\") pod \"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf\" (UID: \"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf\") " Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.535010 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-config-data\") pod \"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf\" (UID: \"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf\") " Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.542071 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-kube-api-access-mc4ns" (OuterVolumeSpecName: "kube-api-access-mc4ns") pod "1e139bd2-3b82-4d07-b96d-6d5b5ca947bf" (UID: "1e139bd2-3b82-4d07-b96d-6d5b5ca947bf"). InnerVolumeSpecName "kube-api-access-mc4ns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.569547 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-config-data" (OuterVolumeSpecName: "config-data") pod "1e139bd2-3b82-4d07-b96d-6d5b5ca947bf" (UID: "1e139bd2-3b82-4d07-b96d-6d5b5ca947bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.583275 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e139bd2-3b82-4d07-b96d-6d5b5ca947bf" (UID: "1e139bd2-3b82-4d07-b96d-6d5b5ca947bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.639183 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.639223 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc4ns\" (UniqueName: \"kubernetes.io/projected/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-kube-api-access-mc4ns\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.639239 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.849534 4739 generic.go:334] "Generic (PLEG): container finished" podID="1e139bd2-3b82-4d07-b96d-6d5b5ca947bf" containerID="7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d" exitCode=0 Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.849589 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.849607 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf","Type":"ContainerDied","Data":"7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d"} Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.850054 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1e139bd2-3b82-4d07-b96d-6d5b5ca947bf","Type":"ContainerDied","Data":"e69781824609bd98645ab94c4295aca272bc4312c159ff634e1d46d63e803b1a"} Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.850106 4739 scope.go:117] "RemoveContainer" containerID="7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.881818 4739 scope.go:117] "RemoveContainer" containerID="7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d" Dec 01 15:55:05 crc kubenswrapper[4739]: E1201 15:55:05.883209 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d\": container with ID starting with 7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d not found: ID does not exist" containerID="7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.883265 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d"} err="failed to get container status \"7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d\": rpc error: code = NotFound desc = could not find container \"7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d\": container with ID starting with 7a06893565b3cd19de11a632ae171c2f0fd371dd659d797196ee3b4b691e142d not found: ID does not exist" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.914547 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.928205 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.941405 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 15:55:05 crc kubenswrapper[4739]: E1201 15:55:05.941829 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e139bd2-3b82-4d07-b96d-6d5b5ca947bf" containerName="nova-scheduler-scheduler" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.941850 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e139bd2-3b82-4d07-b96d-6d5b5ca947bf" containerName="nova-scheduler-scheduler" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.942350 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e139bd2-3b82-4d07-b96d-6d5b5ca947bf" containerName="nova-scheduler-scheduler" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.943291 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.945442 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 15:55:05 crc kubenswrapper[4739]: I1201 15:55:05.951217 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.044754 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a928221-a9c0-4602-81d2-f0dede25dafa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7a928221-a9c0-4602-81d2-f0dede25dafa\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.044807 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a928221-a9c0-4602-81d2-f0dede25dafa-config-data\") pod \"nova-scheduler-0\" (UID: \"7a928221-a9c0-4602-81d2-f0dede25dafa\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.044907 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm8m7\" (UniqueName: \"kubernetes.io/projected/7a928221-a9c0-4602-81d2-f0dede25dafa-kube-api-access-tm8m7\") pod \"nova-scheduler-0\" (UID: \"7a928221-a9c0-4602-81d2-f0dede25dafa\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.113787 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.146818 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a928221-a9c0-4602-81d2-f0dede25dafa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7a928221-a9c0-4602-81d2-f0dede25dafa\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.146866 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a928221-a9c0-4602-81d2-f0dede25dafa-config-data\") pod \"nova-scheduler-0\" (UID: \"7a928221-a9c0-4602-81d2-f0dede25dafa\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.146953 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm8m7\" (UniqueName: \"kubernetes.io/projected/7a928221-a9c0-4602-81d2-f0dede25dafa-kube-api-access-tm8m7\") pod \"nova-scheduler-0\" (UID: \"7a928221-a9c0-4602-81d2-f0dede25dafa\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.156265 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a928221-a9c0-4602-81d2-f0dede25dafa-config-data\") pod \"nova-scheduler-0\" (UID: \"7a928221-a9c0-4602-81d2-f0dede25dafa\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.162132 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a928221-a9c0-4602-81d2-f0dede25dafa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7a928221-a9c0-4602-81d2-f0dede25dafa\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.164692 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm8m7\" (UniqueName: \"kubernetes.io/projected/7a928221-a9c0-4602-81d2-f0dede25dafa-kube-api-access-tm8m7\") pod \"nova-scheduler-0\" (UID: \"7a928221-a9c0-4602-81d2-f0dede25dafa\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.262352 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.490572 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e139bd2-3b82-4d07-b96d-6d5b5ca947bf" path="/var/lib/kubelet/pods/1e139bd2-3b82-4d07-b96d-6d5b5ca947bf/volumes" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.667814 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.708196 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.769552 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6afdc34-badd-4095-aa08-f2350c611f9d-config-data\") pod \"b6afdc34-badd-4095-aa08-f2350c611f9d\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.769611 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6afdc34-badd-4095-aa08-f2350c611f9d-combined-ca-bundle\") pod \"b6afdc34-badd-4095-aa08-f2350c611f9d\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.769694 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6bv5\" (UniqueName: \"kubernetes.io/projected/b6afdc34-badd-4095-aa08-f2350c611f9d-kube-api-access-l6bv5\") pod \"b6afdc34-badd-4095-aa08-f2350c611f9d\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.769724 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6afdc34-badd-4095-aa08-f2350c611f9d-logs\") pod \"b6afdc34-badd-4095-aa08-f2350c611f9d\" (UID: \"b6afdc34-badd-4095-aa08-f2350c611f9d\") " Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.770219 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6afdc34-badd-4095-aa08-f2350c611f9d-logs" (OuterVolumeSpecName: "logs") pod "b6afdc34-badd-4095-aa08-f2350c611f9d" (UID: "b6afdc34-badd-4095-aa08-f2350c611f9d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.774171 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6afdc34-badd-4095-aa08-f2350c611f9d-kube-api-access-l6bv5" (OuterVolumeSpecName: "kube-api-access-l6bv5") pod "b6afdc34-badd-4095-aa08-f2350c611f9d" (UID: "b6afdc34-badd-4095-aa08-f2350c611f9d"). InnerVolumeSpecName "kube-api-access-l6bv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.802682 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6afdc34-badd-4095-aa08-f2350c611f9d-config-data" (OuterVolumeSpecName: "config-data") pod "b6afdc34-badd-4095-aa08-f2350c611f9d" (UID: "b6afdc34-badd-4095-aa08-f2350c611f9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.809982 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6afdc34-badd-4095-aa08-f2350c611f9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6afdc34-badd-4095-aa08-f2350c611f9d" (UID: "b6afdc34-badd-4095-aa08-f2350c611f9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.859251 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7a928221-a9c0-4602-81d2-f0dede25dafa","Type":"ContainerStarted","Data":"ab5dc501fa1784d99cee69fe61ef0e746973eea7aa90ccc86a4f3ae438b7cefa"} Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.862531 4739 generic.go:334] "Generic (PLEG): container finished" podID="b6afdc34-badd-4095-aa08-f2350c611f9d" containerID="d0b2941b270052328077bd733a15a6d7e32636d3f8b755640441e218a1bc7a1c" exitCode=0 Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.862647 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b6afdc34-badd-4095-aa08-f2350c611f9d","Type":"ContainerDied","Data":"d0b2941b270052328077bd733a15a6d7e32636d3f8b755640441e218a1bc7a1c"} Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.862701 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b6afdc34-badd-4095-aa08-f2350c611f9d","Type":"ContainerDied","Data":"591d513ba70da4b468eaca79781d6f6799b99367a93331911c275a4e4c7e699c"} Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.862724 4739 scope.go:117] "RemoveContainer" containerID="d0b2941b270052328077bd733a15a6d7e32636d3f8b755640441e218a1bc7a1c" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.863113 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.873101 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6afdc34-badd-4095-aa08-f2350c611f9d-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.873132 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6afdc34-badd-4095-aa08-f2350c611f9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.873144 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6bv5\" (UniqueName: \"kubernetes.io/projected/b6afdc34-badd-4095-aa08-f2350c611f9d-kube-api-access-l6bv5\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.873154 4739 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6afdc34-badd-4095-aa08-f2350c611f9d-logs\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.888825 4739 scope.go:117] "RemoveContainer" containerID="2ab24ed3341c32b08df4ea109ef756de9b0ae9f562035ae1631724747467ce78" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.926764 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.941162 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.942632 4739 scope.go:117] "RemoveContainer" containerID="d0b2941b270052328077bd733a15a6d7e32636d3f8b755640441e218a1bc7a1c" Dec 01 15:55:06 crc kubenswrapper[4739]: E1201 15:55:06.943911 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0b2941b270052328077bd733a15a6d7e32636d3f8b755640441e218a1bc7a1c\": container with ID starting with d0b2941b270052328077bd733a15a6d7e32636d3f8b755640441e218a1bc7a1c not found: ID does not exist" containerID="d0b2941b270052328077bd733a15a6d7e32636d3f8b755640441e218a1bc7a1c" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.943942 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0b2941b270052328077bd733a15a6d7e32636d3f8b755640441e218a1bc7a1c"} err="failed to get container status \"d0b2941b270052328077bd733a15a6d7e32636d3f8b755640441e218a1bc7a1c\": rpc error: code = NotFound desc = could not find container \"d0b2941b270052328077bd733a15a6d7e32636d3f8b755640441e218a1bc7a1c\": container with ID starting with d0b2941b270052328077bd733a15a6d7e32636d3f8b755640441e218a1bc7a1c not found: ID does not exist" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.943962 4739 scope.go:117] "RemoveContainer" containerID="2ab24ed3341c32b08df4ea109ef756de9b0ae9f562035ae1631724747467ce78" Dec 01 15:55:06 crc kubenswrapper[4739]: E1201 15:55:06.944166 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ab24ed3341c32b08df4ea109ef756de9b0ae9f562035ae1631724747467ce78\": container with ID starting with 2ab24ed3341c32b08df4ea109ef756de9b0ae9f562035ae1631724747467ce78 not found: ID does not exist" containerID="2ab24ed3341c32b08df4ea109ef756de9b0ae9f562035ae1631724747467ce78" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.944188 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ab24ed3341c32b08df4ea109ef756de9b0ae9f562035ae1631724747467ce78"} err="failed to get container status \"2ab24ed3341c32b08df4ea109ef756de9b0ae9f562035ae1631724747467ce78\": rpc error: code = NotFound desc = could not find container \"2ab24ed3341c32b08df4ea109ef756de9b0ae9f562035ae1631724747467ce78\": container with ID starting with 2ab24ed3341c32b08df4ea109ef756de9b0ae9f562035ae1631724747467ce78 not found: ID does not exist" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.952966 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:06 crc kubenswrapper[4739]: E1201 15:55:06.953550 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6afdc34-badd-4095-aa08-f2350c611f9d" containerName="nova-api-log" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.953622 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6afdc34-badd-4095-aa08-f2350c611f9d" containerName="nova-api-log" Dec 01 15:55:06 crc kubenswrapper[4739]: E1201 15:55:06.953697 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6afdc34-badd-4095-aa08-f2350c611f9d" containerName="nova-api-api" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.953771 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6afdc34-badd-4095-aa08-f2350c611f9d" containerName="nova-api-api" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.954069 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6afdc34-badd-4095-aa08-f2350c611f9d" containerName="nova-api-log" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.954175 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6afdc34-badd-4095-aa08-f2350c611f9d" containerName="nova-api-api" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.955337 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.958682 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 15:55:06 crc kubenswrapper[4739]: I1201 15:55:06.975801 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.077676 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ed721c3-b669-4601-a166-d851e7960b2c-logs\") pod \"nova-api-0\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " pod="openstack/nova-api-0" Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.077726 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ed721c3-b669-4601-a166-d851e7960b2c-config-data\") pod \"nova-api-0\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " pod="openstack/nova-api-0" Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.077845 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ed721c3-b669-4601-a166-d851e7960b2c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " pod="openstack/nova-api-0" Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.077875 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f48fl\" (UniqueName: \"kubernetes.io/projected/7ed721c3-b669-4601-a166-d851e7960b2c-kube-api-access-f48fl\") pod \"nova-api-0\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " pod="openstack/nova-api-0" Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.178993 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ed721c3-b669-4601-a166-d851e7960b2c-logs\") pod \"nova-api-0\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " pod="openstack/nova-api-0" Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.179412 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ed721c3-b669-4601-a166-d851e7960b2c-config-data\") pod \"nova-api-0\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " pod="openstack/nova-api-0" Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.179556 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ed721c3-b669-4601-a166-d851e7960b2c-logs\") pod \"nova-api-0\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " pod="openstack/nova-api-0" Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.179759 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ed721c3-b669-4601-a166-d851e7960b2c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " pod="openstack/nova-api-0" Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.179902 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f48fl\" (UniqueName: \"kubernetes.io/projected/7ed721c3-b669-4601-a166-d851e7960b2c-kube-api-access-f48fl\") pod \"nova-api-0\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " pod="openstack/nova-api-0" Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.180058 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.180945 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.183051 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ed721c3-b669-4601-a166-d851e7960b2c-config-data\") pod \"nova-api-0\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " pod="openstack/nova-api-0" Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.183224 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ed721c3-b669-4601-a166-d851e7960b2c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " pod="openstack/nova-api-0" Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.208307 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f48fl\" (UniqueName: \"kubernetes.io/projected/7ed721c3-b669-4601-a166-d851e7960b2c-kube-api-access-f48fl\") pod \"nova-api-0\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " pod="openstack/nova-api-0" Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.278068 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.720315 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.871066 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ed721c3-b669-4601-a166-d851e7960b2c","Type":"ContainerStarted","Data":"7da751400ea2fdde0fab17feff61a00ce2d72d6321b14542df24c9c64e538921"} Dec 01 15:55:07 crc kubenswrapper[4739]: I1201 15:55:07.872484 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7a928221-a9c0-4602-81d2-f0dede25dafa","Type":"ContainerStarted","Data":"ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d"} Dec 01 15:55:08 crc kubenswrapper[4739]: I1201 15:55:08.493328 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6afdc34-badd-4095-aa08-f2350c611f9d" path="/var/lib/kubelet/pods/b6afdc34-badd-4095-aa08-f2350c611f9d/volumes" Dec 01 15:55:08 crc kubenswrapper[4739]: I1201 15:55:08.519655 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.519631787 podStartE2EDuration="3.519631787s" podCreationTimestamp="2025-12-01 15:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:55:07.886565543 +0000 UTC m=+1209.712311647" watchObservedRunningTime="2025-12-01 15:55:08.519631787 +0000 UTC m=+1210.345377891" Dec 01 15:55:08 crc kubenswrapper[4739]: I1201 15:55:08.897501 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ed721c3-b669-4601-a166-d851e7960b2c","Type":"ContainerStarted","Data":"9ebc3a7b9cabc58ccb10e9b42226d4467672ac309fcc8cd15f09e88838831068"} Dec 01 15:55:08 crc kubenswrapper[4739]: I1201 15:55:08.899265 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ed721c3-b669-4601-a166-d851e7960b2c","Type":"ContainerStarted","Data":"833e1e56711e5b5c55b51087caee82082231fc03077d4f3b52b9c5579c45ccc1"} Dec 01 15:55:08 crc kubenswrapper[4739]: I1201 15:55:08.935481 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.935454884 podStartE2EDuration="2.935454884s" podCreationTimestamp="2025-12-01 15:55:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:55:08.931550833 +0000 UTC m=+1210.757296957" watchObservedRunningTime="2025-12-01 15:55:08.935454884 +0000 UTC m=+1210.761200988" Dec 01 15:55:11 crc kubenswrapper[4739]: I1201 15:55:11.263393 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 15:55:12 crc kubenswrapper[4739]: I1201 15:55:12.180941 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 15:55:12 crc kubenswrapper[4739]: I1201 15:55:12.181020 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 15:55:12 crc kubenswrapper[4739]: I1201 15:55:12.840864 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 15:55:13 crc kubenswrapper[4739]: I1201 15:55:13.199646 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b30c462a-2b55-4187-931e-12e78990741a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.175:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 15:55:13 crc kubenswrapper[4739]: I1201 15:55:13.199680 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b30c462a-2b55-4187-931e-12e78990741a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.175:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 15:55:15 crc kubenswrapper[4739]: I1201 15:55:15.507696 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 15:55:15 crc kubenswrapper[4739]: I1201 15:55:15.510364 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="51f0a0d6-a82c-495d-9cd4-141c824e6453" containerName="kube-state-metrics" containerID="cri-o://370345a48750c51d8deaa0f4eebdbea37cf4b8ea971f93c0a9e129e6f3f76995" gracePeriod=30 Dec 01 15:55:15 crc kubenswrapper[4739]: I1201 15:55:15.973406 4739 generic.go:334] "Generic (PLEG): container finished" podID="51f0a0d6-a82c-495d-9cd4-141c824e6453" containerID="370345a48750c51d8deaa0f4eebdbea37cf4b8ea971f93c0a9e129e6f3f76995" exitCode=2 Dec 01 15:55:15 crc kubenswrapper[4739]: I1201 15:55:15.973801 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"51f0a0d6-a82c-495d-9cd4-141c824e6453","Type":"ContainerDied","Data":"370345a48750c51d8deaa0f4eebdbea37cf4b8ea971f93c0a9e129e6f3f76995"} Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.137350 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.256028 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldx8c\" (UniqueName: \"kubernetes.io/projected/51f0a0d6-a82c-495d-9cd4-141c824e6453-kube-api-access-ldx8c\") pod \"51f0a0d6-a82c-495d-9cd4-141c824e6453\" (UID: \"51f0a0d6-a82c-495d-9cd4-141c824e6453\") " Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.261857 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51f0a0d6-a82c-495d-9cd4-141c824e6453-kube-api-access-ldx8c" (OuterVolumeSpecName: "kube-api-access-ldx8c") pod "51f0a0d6-a82c-495d-9cd4-141c824e6453" (UID: "51f0a0d6-a82c-495d-9cd4-141c824e6453"). InnerVolumeSpecName "kube-api-access-ldx8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.262872 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.291902 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.358616 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldx8c\" (UniqueName: \"kubernetes.io/projected/51f0a0d6-a82c-495d-9cd4-141c824e6453-kube-api-access-ldx8c\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.632677 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.633921 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="sg-core" containerID="cri-o://4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0" gracePeriod=30 Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.634058 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="proxy-httpd" containerID="cri-o://9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0" gracePeriod=30 Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.634142 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="ceilometer-notification-agent" containerID="cri-o://73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560" gracePeriod=30 Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.633811 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="ceilometer-central-agent" containerID="cri-o://774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df" gracePeriod=30 Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.985085 4739 generic.go:334] "Generic (PLEG): container finished" podID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerID="9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0" exitCode=0 Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.985602 4739 generic.go:334] "Generic (PLEG): container finished" podID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerID="4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0" exitCode=2 Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.985152 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33904e6b-d018-4a30-a9c4-f57657f1795c","Type":"ContainerDied","Data":"9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0"} Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.985938 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33904e6b-d018-4a30-a9c4-f57657f1795c","Type":"ContainerDied","Data":"4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0"} Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.987513 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"51f0a0d6-a82c-495d-9cd4-141c824e6453","Type":"ContainerDied","Data":"b34193f93ed8c47508313827e35ccc38d972ecb418051376c9b13caa2420cf72"} Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.987548 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 15:55:16 crc kubenswrapper[4739]: I1201 15:55:16.987571 4739 scope.go:117] "RemoveContainer" containerID="370345a48750c51d8deaa0f4eebdbea37cf4b8ea971f93c0a9e129e6f3f76995" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.013237 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.022044 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.037114 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.060578 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 15:55:17 crc kubenswrapper[4739]: E1201 15:55:17.061982 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51f0a0d6-a82c-495d-9cd4-141c824e6453" containerName="kube-state-metrics" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.062023 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="51f0a0d6-a82c-495d-9cd4-141c824e6453" containerName="kube-state-metrics" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.063827 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="51f0a0d6-a82c-495d-9cd4-141c824e6453" containerName="kube-state-metrics" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.064710 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.067377 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.068019 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.087649 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.181584 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e3c95a-b796-44da-93fe-e958ec7ae3c1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"85e3c95a-b796-44da-93fe-e958ec7ae3c1\") " pod="openstack/kube-state-metrics-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.181814 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/85e3c95a-b796-44da-93fe-e958ec7ae3c1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"85e3c95a-b796-44da-93fe-e958ec7ae3c1\") " pod="openstack/kube-state-metrics-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.182011 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mvhv\" (UniqueName: \"kubernetes.io/projected/85e3c95a-b796-44da-93fe-e958ec7ae3c1-kube-api-access-9mvhv\") pod \"kube-state-metrics-0\" (UID: \"85e3c95a-b796-44da-93fe-e958ec7ae3c1\") " pod="openstack/kube-state-metrics-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.182200 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/85e3c95a-b796-44da-93fe-e958ec7ae3c1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"85e3c95a-b796-44da-93fe-e958ec7ae3c1\") " pod="openstack/kube-state-metrics-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.279043 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.279115 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.288478 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mvhv\" (UniqueName: \"kubernetes.io/projected/85e3c95a-b796-44da-93fe-e958ec7ae3c1-kube-api-access-9mvhv\") pod \"kube-state-metrics-0\" (UID: \"85e3c95a-b796-44da-93fe-e958ec7ae3c1\") " pod="openstack/kube-state-metrics-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.288600 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/85e3c95a-b796-44da-93fe-e958ec7ae3c1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"85e3c95a-b796-44da-93fe-e958ec7ae3c1\") " pod="openstack/kube-state-metrics-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.288658 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e3c95a-b796-44da-93fe-e958ec7ae3c1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"85e3c95a-b796-44da-93fe-e958ec7ae3c1\") " pod="openstack/kube-state-metrics-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.288769 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/85e3c95a-b796-44da-93fe-e958ec7ae3c1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"85e3c95a-b796-44da-93fe-e958ec7ae3c1\") " pod="openstack/kube-state-metrics-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.302351 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/85e3c95a-b796-44da-93fe-e958ec7ae3c1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"85e3c95a-b796-44da-93fe-e958ec7ae3c1\") " pod="openstack/kube-state-metrics-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.302516 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e3c95a-b796-44da-93fe-e958ec7ae3c1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"85e3c95a-b796-44da-93fe-e958ec7ae3c1\") " pod="openstack/kube-state-metrics-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.302836 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/85e3c95a-b796-44da-93fe-e958ec7ae3c1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"85e3c95a-b796-44da-93fe-e958ec7ae3c1\") " pod="openstack/kube-state-metrics-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.321582 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mvhv\" (UniqueName: \"kubernetes.io/projected/85e3c95a-b796-44da-93fe-e958ec7ae3c1-kube-api-access-9mvhv\") pod \"kube-state-metrics-0\" (UID: \"85e3c95a-b796-44da-93fe-e958ec7ae3c1\") " pod="openstack/kube-state-metrics-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.395582 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.439156 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.492882 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-scripts\") pod \"33904e6b-d018-4a30-a9c4-f57657f1795c\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.492937 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-sg-core-conf-yaml\") pod \"33904e6b-d018-4a30-a9c4-f57657f1795c\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.492992 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t245v\" (UniqueName: \"kubernetes.io/projected/33904e6b-d018-4a30-a9c4-f57657f1795c-kube-api-access-t245v\") pod \"33904e6b-d018-4a30-a9c4-f57657f1795c\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.493053 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33904e6b-d018-4a30-a9c4-f57657f1795c-run-httpd\") pod \"33904e6b-d018-4a30-a9c4-f57657f1795c\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.493082 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-combined-ca-bundle\") pod \"33904e6b-d018-4a30-a9c4-f57657f1795c\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.493110 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-config-data\") pod \"33904e6b-d018-4a30-a9c4-f57657f1795c\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.493227 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33904e6b-d018-4a30-a9c4-f57657f1795c-log-httpd\") pod \"33904e6b-d018-4a30-a9c4-f57657f1795c\" (UID: \"33904e6b-d018-4a30-a9c4-f57657f1795c\") " Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.494167 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33904e6b-d018-4a30-a9c4-f57657f1795c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "33904e6b-d018-4a30-a9c4-f57657f1795c" (UID: "33904e6b-d018-4a30-a9c4-f57657f1795c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.493916 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33904e6b-d018-4a30-a9c4-f57657f1795c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "33904e6b-d018-4a30-a9c4-f57657f1795c" (UID: "33904e6b-d018-4a30-a9c4-f57657f1795c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.496511 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-scripts" (OuterVolumeSpecName: "scripts") pod "33904e6b-d018-4a30-a9c4-f57657f1795c" (UID: "33904e6b-d018-4a30-a9c4-f57657f1795c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.496588 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33904e6b-d018-4a30-a9c4-f57657f1795c-kube-api-access-t245v" (OuterVolumeSpecName: "kube-api-access-t245v") pod "33904e6b-d018-4a30-a9c4-f57657f1795c" (UID: "33904e6b-d018-4a30-a9c4-f57657f1795c"). InnerVolumeSpecName "kube-api-access-t245v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.541533 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "33904e6b-d018-4a30-a9c4-f57657f1795c" (UID: "33904e6b-d018-4a30-a9c4-f57657f1795c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.575999 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33904e6b-d018-4a30-a9c4-f57657f1795c" (UID: "33904e6b-d018-4a30-a9c4-f57657f1795c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.595756 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.595799 4739 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.595814 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t245v\" (UniqueName: \"kubernetes.io/projected/33904e6b-d018-4a30-a9c4-f57657f1795c-kube-api-access-t245v\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.595827 4739 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33904e6b-d018-4a30-a9c4-f57657f1795c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.595842 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.595854 4739 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33904e6b-d018-4a30-a9c4-f57657f1795c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.641713 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-config-data" (OuterVolumeSpecName: "config-data") pod "33904e6b-d018-4a30-a9c4-f57657f1795c" (UID: "33904e6b-d018-4a30-a9c4-f57657f1795c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.698897 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33904e6b-d018-4a30-a9c4-f57657f1795c-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:17 crc kubenswrapper[4739]: I1201 15:55:17.901610 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.000091 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"85e3c95a-b796-44da-93fe-e958ec7ae3c1","Type":"ContainerStarted","Data":"583f49375a5ef13108b569c39535f4ebf886655067e6a29114443a6a7618cf8a"} Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.004670 4739 generic.go:334] "Generic (PLEG): container finished" podID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerID="73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560" exitCode=0 Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.004704 4739 generic.go:334] "Generic (PLEG): container finished" podID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerID="774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df" exitCode=0 Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.004742 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33904e6b-d018-4a30-a9c4-f57657f1795c","Type":"ContainerDied","Data":"73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560"} Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.004768 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33904e6b-d018-4a30-a9c4-f57657f1795c","Type":"ContainerDied","Data":"774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df"} Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.004777 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33904e6b-d018-4a30-a9c4-f57657f1795c","Type":"ContainerDied","Data":"995329ae8eb8a79e32d2310a95c4ff4383ef4d895dce629d586cedd7e8ab20a0"} Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.004792 4739 scope.go:117] "RemoveContainer" containerID="9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.004913 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.041785 4739 scope.go:117] "RemoveContainer" containerID="4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.050288 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.063774 4739 scope.go:117] "RemoveContainer" containerID="73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.065541 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.080518 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:55:18 crc kubenswrapper[4739]: E1201 15:55:18.081105 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="ceilometer-notification-agent" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.081136 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="ceilometer-notification-agent" Dec 01 15:55:18 crc kubenswrapper[4739]: E1201 15:55:18.081174 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="ceilometer-central-agent" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.081187 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="ceilometer-central-agent" Dec 01 15:55:18 crc kubenswrapper[4739]: E1201 15:55:18.081224 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="proxy-httpd" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.081236 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="proxy-httpd" Dec 01 15:55:18 crc kubenswrapper[4739]: E1201 15:55:18.081257 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="sg-core" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.081269 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="sg-core" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.081572 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="ceilometer-central-agent" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.081612 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="proxy-httpd" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.081633 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="ceilometer-notification-agent" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.081669 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" containerName="sg-core" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.084858 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.088492 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.089359 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.089711 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.089836 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.101931 4739 scope.go:117] "RemoveContainer" containerID="774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.130366 4739 scope.go:117] "RemoveContainer" containerID="9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0" Dec 01 15:55:18 crc kubenswrapper[4739]: E1201 15:55:18.130755 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0\": container with ID starting with 9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0 not found: ID does not exist" containerID="9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.130783 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0"} err="failed to get container status \"9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0\": rpc error: code = NotFound desc = could not find container \"9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0\": container with ID starting with 9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0 not found: ID does not exist" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.130805 4739 scope.go:117] "RemoveContainer" containerID="4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0" Dec 01 15:55:18 crc kubenswrapper[4739]: E1201 15:55:18.131016 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0\": container with ID starting with 4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0 not found: ID does not exist" containerID="4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.131037 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0"} err="failed to get container status \"4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0\": rpc error: code = NotFound desc = could not find container \"4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0\": container with ID starting with 4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0 not found: ID does not exist" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.131049 4739 scope.go:117] "RemoveContainer" containerID="73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560" Dec 01 15:55:18 crc kubenswrapper[4739]: E1201 15:55:18.131220 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560\": container with ID starting with 73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560 not found: ID does not exist" containerID="73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.131236 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560"} err="failed to get container status \"73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560\": rpc error: code = NotFound desc = could not find container \"73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560\": container with ID starting with 73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560 not found: ID does not exist" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.131247 4739 scope.go:117] "RemoveContainer" containerID="774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df" Dec 01 15:55:18 crc kubenswrapper[4739]: E1201 15:55:18.131405 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df\": container with ID starting with 774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df not found: ID does not exist" containerID="774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.131491 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df"} err="failed to get container status \"774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df\": rpc error: code = NotFound desc = could not find container \"774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df\": container with ID starting with 774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df not found: ID does not exist" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.131507 4739 scope.go:117] "RemoveContainer" containerID="9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.132269 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0"} err="failed to get container status \"9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0\": rpc error: code = NotFound desc = could not find container \"9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0\": container with ID starting with 9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0 not found: ID does not exist" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.132290 4739 scope.go:117] "RemoveContainer" containerID="4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.132638 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0"} err="failed to get container status \"4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0\": rpc error: code = NotFound desc = could not find container \"4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0\": container with ID starting with 4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0 not found: ID does not exist" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.132671 4739 scope.go:117] "RemoveContainer" containerID="73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.132927 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560"} err="failed to get container status \"73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560\": rpc error: code = NotFound desc = could not find container \"73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560\": container with ID starting with 73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560 not found: ID does not exist" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.132947 4739 scope.go:117] "RemoveContainer" containerID="774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.133170 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df"} err="failed to get container status \"774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df\": rpc error: code = NotFound desc = could not find container \"774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df\": container with ID starting with 774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df not found: ID does not exist" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.214656 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb4aa436-d5da-4b15-b0d5-dcad7118be13-run-httpd\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.214711 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.214739 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsg88\" (UniqueName: \"kubernetes.io/projected/fb4aa436-d5da-4b15-b0d5-dcad7118be13-kube-api-access-rsg88\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.214771 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-scripts\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.214876 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.215074 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-config-data\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.215185 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb4aa436-d5da-4b15-b0d5-dcad7118be13-log-httpd\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.215218 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.317259 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-config-data\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.317595 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb4aa436-d5da-4b15-b0d5-dcad7118be13-log-httpd\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.317618 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.317678 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb4aa436-d5da-4b15-b0d5-dcad7118be13-run-httpd\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.317708 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.317744 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsg88\" (UniqueName: \"kubernetes.io/projected/fb4aa436-d5da-4b15-b0d5-dcad7118be13-kube-api-access-rsg88\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.317782 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-scripts\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.317817 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.318119 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb4aa436-d5da-4b15-b0d5-dcad7118be13-run-httpd\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.318397 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb4aa436-d5da-4b15-b0d5-dcad7118be13-log-httpd\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.320557 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7ed721c3-b669-4601-a166-d851e7960b2c" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.177:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.320558 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7ed721c3-b669-4601-a166-d851e7960b2c" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.177:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.323098 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.324164 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-scripts\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.324467 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-config-data\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.327910 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.339924 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsg88\" (UniqueName: \"kubernetes.io/projected/fb4aa436-d5da-4b15-b0d5-dcad7118be13-kube-api-access-rsg88\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.364004 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.407668 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.495803 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33904e6b-d018-4a30-a9c4-f57657f1795c" path="/var/lib/kubelet/pods/33904e6b-d018-4a30-a9c4-f57657f1795c/volumes" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.509015 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51f0a0d6-a82c-495d-9cd4-141c824e6453" path="/var/lib/kubelet/pods/51f0a0d6-a82c-495d-9cd4-141c824e6453/volumes" Dec 01 15:55:18 crc kubenswrapper[4739]: I1201 15:55:18.853429 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:55:19 crc kubenswrapper[4739]: I1201 15:55:19.028453 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb4aa436-d5da-4b15-b0d5-dcad7118be13","Type":"ContainerStarted","Data":"67ac5e1768e476ec7c75f48fce936ef9e56d1829281e8f40031fafda107a4b4b"} Dec 01 15:55:19 crc kubenswrapper[4739]: I1201 15:55:19.029953 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"85e3c95a-b796-44da-93fe-e958ec7ae3c1","Type":"ContainerStarted","Data":"d65bd9d99348b8941d808b64d3098a3c6a986f0df4c3312f1e021c9f4fe033f9"} Dec 01 15:55:19 crc kubenswrapper[4739]: I1201 15:55:19.029993 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 15:55:19 crc kubenswrapper[4739]: I1201 15:55:19.051603 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.630732712 podStartE2EDuration="2.051585264s" podCreationTimestamp="2025-12-01 15:55:17 +0000 UTC" firstStartedPulling="2025-12-01 15:55:17.920447856 +0000 UTC m=+1219.746193950" lastFinishedPulling="2025-12-01 15:55:18.341300408 +0000 UTC m=+1220.167046502" observedRunningTime="2025-12-01 15:55:19.046254928 +0000 UTC m=+1220.872001042" watchObservedRunningTime="2025-12-01 15:55:19.051585264 +0000 UTC m=+1220.877331358" Dec 01 15:55:20 crc kubenswrapper[4739]: I1201 15:55:20.050562 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb4aa436-d5da-4b15-b0d5-dcad7118be13","Type":"ContainerStarted","Data":"d660793640981a88857238007f7dc9dbc3b55c7fe2d3a50eaeacb8d5829c8555"} Dec 01 15:55:21 crc kubenswrapper[4739]: I1201 15:55:21.067259 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb4aa436-d5da-4b15-b0d5-dcad7118be13","Type":"ContainerStarted","Data":"eef22ca6c668d6f576f86a6e3d69bc496172a0250fcf9a88cc4dbe3962bd46f6"} Dec 01 15:55:22 crc kubenswrapper[4739]: I1201 15:55:22.080660 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb4aa436-d5da-4b15-b0d5-dcad7118be13","Type":"ContainerStarted","Data":"fedb362e102f965cec0fb488edbee589d98f7594016a736e34fe4f0c29698e42"} Dec 01 15:55:22 crc kubenswrapper[4739]: I1201 15:55:22.185335 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 15:55:22 crc kubenswrapper[4739]: I1201 15:55:22.191769 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 15:55:22 crc kubenswrapper[4739]: I1201 15:55:22.192684 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 15:55:23 crc kubenswrapper[4739]: I1201 15:55:23.096481 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb4aa436-d5da-4b15-b0d5-dcad7118be13","Type":"ContainerStarted","Data":"2fce24f366ea490a132aca38413502b104477bb84c102c3bab3b06fc5b645398"} Dec 01 15:55:23 crc kubenswrapper[4739]: I1201 15:55:23.106574 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 15:55:23 crc kubenswrapper[4739]: I1201 15:55:23.121000 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.3723024750000001 podStartE2EDuration="5.120979222s" podCreationTimestamp="2025-12-01 15:55:18 +0000 UTC" firstStartedPulling="2025-12-01 15:55:18.859877417 +0000 UTC m=+1220.685623511" lastFinishedPulling="2025-12-01 15:55:22.608554134 +0000 UTC m=+1224.434300258" observedRunningTime="2025-12-01 15:55:23.119018152 +0000 UTC m=+1224.944764276" watchObservedRunningTime="2025-12-01 15:55:23.120979222 +0000 UTC m=+1224.946725326" Dec 01 15:55:24 crc kubenswrapper[4739]: I1201 15:55:24.108476 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 15:55:24 crc kubenswrapper[4739]: E1201 15:55:24.639586 4739 manager.go:1116] Failed to create existing container: /kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33904e6b_d018_4a30_a9c4_f57657f1795c.slice/crio-995329ae8eb8a79e32d2310a95c4ff4383ef4d895dce629d586cedd7e8ab20a0: Error finding container 995329ae8eb8a79e32d2310a95c4ff4383ef4d895dce629d586cedd7e8ab20a0: Status 404 returned error can't find the container with id 995329ae8eb8a79e32d2310a95c4ff4383ef4d895dce629d586cedd7e8ab20a0 Dec 01 15:55:24 crc kubenswrapper[4739]: E1201 15:55:24.882358 4739 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33904e6b_d018_4a30_a9c4_f57657f1795c.slice/crio-conmon-4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51f0a0d6_a82c_495d_9cd4_141c824e6453.slice/crio-b34193f93ed8c47508313827e35ccc38d972ecb418051376c9b13caa2420cf72\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33904e6b_d018_4a30_a9c4_f57657f1795c.slice/crio-73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33904e6b_d018_4a30_a9c4_f57657f1795c.slice/crio-conmon-9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33904e6b_d018_4a30_a9c4_f57657f1795c.slice/crio-conmon-73ff26daeaa98a00cecdfb2bde33b79997afceb9ea906b58c89ca83006084560.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33904e6b_d018_4a30_a9c4_f57657f1795c.slice/crio-774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d9292e1_fbbf_46d2_85ad_4eda18061dbb.slice/crio-conmon-0b432b1274db4acbebc79ac56068c6a2a2fc8a58426387136bdb5c2b3a828ab2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33904e6b_d018_4a30_a9c4_f57657f1795c.slice/crio-4606b8156c9ff0ef177c3e4eb72b3a720e5bfc7621344351a290ef3acaf212b0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33904e6b_d018_4a30_a9c4_f57657f1795c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51f0a0d6_a82c_495d_9cd4_141c824e6453.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51f0a0d6_a82c_495d_9cd4_141c824e6453.slice/crio-370345a48750c51d8deaa0f4eebdbea37cf4b8ea971f93c0a9e129e6f3f76995.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51f0a0d6_a82c_495d_9cd4_141c824e6453.slice/crio-conmon-370345a48750c51d8deaa0f4eebdbea37cf4b8ea971f93c0a9e129e6f3f76995.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33904e6b_d018_4a30_a9c4_f57657f1795c.slice/crio-conmon-774a80564ae254690abeddccace4d8b83af54662ba8a4761c7661b5a6734d7df.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33904e6b_d018_4a30_a9c4_f57657f1795c.slice/crio-9f966024d81d0a448298c5efb421679b988e96355ca8d648eb4d6ae84aafeca0.scope\": RecentStats: unable to find data in memory cache]" Dec 01 15:55:25 crc kubenswrapper[4739]: I1201 15:55:25.119724 4739 generic.go:334] "Generic (PLEG): container finished" podID="7d9292e1-fbbf-46d2-85ad-4eda18061dbb" containerID="0b432b1274db4acbebc79ac56068c6a2a2fc8a58426387136bdb5c2b3a828ab2" exitCode=137 Dec 01 15:55:25 crc kubenswrapper[4739]: I1201 15:55:25.119786 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7d9292e1-fbbf-46d2-85ad-4eda18061dbb","Type":"ContainerDied","Data":"0b432b1274db4acbebc79ac56068c6a2a2fc8a58426387136bdb5c2b3a828ab2"} Dec 01 15:55:25 crc kubenswrapper[4739]: I1201 15:55:25.120126 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7d9292e1-fbbf-46d2-85ad-4eda18061dbb","Type":"ContainerDied","Data":"1da04fcb50be5c944f5302f8c71eac4662e15ca1ed533e97f7baa5e620bb211b"} Dec 01 15:55:25 crc kubenswrapper[4739]: I1201 15:55:25.120162 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1da04fcb50be5c944f5302f8c71eac4662e15ca1ed533e97f7baa5e620bb211b" Dec 01 15:55:25 crc kubenswrapper[4739]: I1201 15:55:25.194669 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:25 crc kubenswrapper[4739]: I1201 15:55:25.250482 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-config-data\") pod \"7d9292e1-fbbf-46d2-85ad-4eda18061dbb\" (UID: \"7d9292e1-fbbf-46d2-85ad-4eda18061dbb\") " Dec 01 15:55:25 crc kubenswrapper[4739]: I1201 15:55:25.250559 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pklg\" (UniqueName: \"kubernetes.io/projected/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-kube-api-access-2pklg\") pod \"7d9292e1-fbbf-46d2-85ad-4eda18061dbb\" (UID: \"7d9292e1-fbbf-46d2-85ad-4eda18061dbb\") " Dec 01 15:55:25 crc kubenswrapper[4739]: I1201 15:55:25.250717 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-combined-ca-bundle\") pod \"7d9292e1-fbbf-46d2-85ad-4eda18061dbb\" (UID: \"7d9292e1-fbbf-46d2-85ad-4eda18061dbb\") " Dec 01 15:55:25 crc kubenswrapper[4739]: I1201 15:55:25.257979 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-kube-api-access-2pklg" (OuterVolumeSpecName: "kube-api-access-2pklg") pod "7d9292e1-fbbf-46d2-85ad-4eda18061dbb" (UID: "7d9292e1-fbbf-46d2-85ad-4eda18061dbb"). InnerVolumeSpecName "kube-api-access-2pklg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:25 crc kubenswrapper[4739]: I1201 15:55:25.285723 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d9292e1-fbbf-46d2-85ad-4eda18061dbb" (UID: "7d9292e1-fbbf-46d2-85ad-4eda18061dbb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:25 crc kubenswrapper[4739]: I1201 15:55:25.303526 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-config-data" (OuterVolumeSpecName: "config-data") pod "7d9292e1-fbbf-46d2-85ad-4eda18061dbb" (UID: "7d9292e1-fbbf-46d2-85ad-4eda18061dbb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:25 crc kubenswrapper[4739]: I1201 15:55:25.354842 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:25 crc kubenswrapper[4739]: I1201 15:55:25.354889 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:25 crc kubenswrapper[4739]: I1201 15:55:25.354910 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pklg\" (UniqueName: \"kubernetes.io/projected/7d9292e1-fbbf-46d2-85ad-4eda18061dbb-kube-api-access-2pklg\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.131658 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.220013 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.240866 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.255386 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 15:55:26 crc kubenswrapper[4739]: E1201 15:55:26.255838 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d9292e1-fbbf-46d2-85ad-4eda18061dbb" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.255864 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d9292e1-fbbf-46d2-85ad-4eda18061dbb" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.256125 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d9292e1-fbbf-46d2-85ad-4eda18061dbb" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.256934 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.260523 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.260730 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.261127 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.266436 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.379725 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.379828 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z4xd\" (UniqueName: \"kubernetes.io/projected/ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0-kube-api-access-6z4xd\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.379879 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.379909 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.380041 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.488370 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z4xd\" (UniqueName: \"kubernetes.io/projected/ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0-kube-api-access-6z4xd\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.488451 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.488476 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.488545 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.488853 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.494120 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.503275 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.510623 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d9292e1-fbbf-46d2-85ad-4eda18061dbb" path="/var/lib/kubelet/pods/7d9292e1-fbbf-46d2-85ad-4eda18061dbb/volumes" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.517721 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.517754 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.523456 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z4xd\" (UniqueName: \"kubernetes.io/projected/ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0-kube-api-access-6z4xd\") pod \"nova-cell1-novncproxy-0\" (UID: \"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:26 crc kubenswrapper[4739]: I1201 15:55:26.581757 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:27 crc kubenswrapper[4739]: W1201 15:55:27.070408 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab4c4993_67b7_46f2_8aeb_0173ffc9d1a0.slice/crio-6251c84b9ce72d0debf9a886d46e309fbc85ec252d98d34da6236364a8f25983 WatchSource:0}: Error finding container 6251c84b9ce72d0debf9a886d46e309fbc85ec252d98d34da6236364a8f25983: Status 404 returned error can't find the container with id 6251c84b9ce72d0debf9a886d46e309fbc85ec252d98d34da6236364a8f25983 Dec 01 15:55:27 crc kubenswrapper[4739]: I1201 15:55:27.074391 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 15:55:27 crc kubenswrapper[4739]: I1201 15:55:27.141405 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0","Type":"ContainerStarted","Data":"6251c84b9ce72d0debf9a886d46e309fbc85ec252d98d34da6236364a8f25983"} Dec 01 15:55:27 crc kubenswrapper[4739]: I1201 15:55:27.287454 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 15:55:27 crc kubenswrapper[4739]: I1201 15:55:27.288784 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 15:55:27 crc kubenswrapper[4739]: I1201 15:55:27.289043 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 15:55:27 crc kubenswrapper[4739]: I1201 15:55:27.292504 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 15:55:27 crc kubenswrapper[4739]: I1201 15:55:27.451009 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.151907 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0","Type":"ContainerStarted","Data":"37c512715c80ae2f5f80a76e684ca4740e0479f8b64d0b5fccffc6eedfe1231a"} Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.152535 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.158388 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.186738 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.186708664 podStartE2EDuration="2.186708664s" podCreationTimestamp="2025-12-01 15:55:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:55:28.17167778 +0000 UTC m=+1229.997423874" watchObservedRunningTime="2025-12-01 15:55:28.186708664 +0000 UTC m=+1230.012454778" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.356454 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-9dldl"] Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.357851 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.384579 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-9dldl"] Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.432885 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smlh4\" (UniqueName: \"kubernetes.io/projected/501d57ed-62bc-452a-b0ae-12835a26cd67-kube-api-access-smlh4\") pod \"dnsmasq-dns-5b856c5697-9dldl\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.432948 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-9dldl\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.432978 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-config\") pod \"dnsmasq-dns-5b856c5697-9dldl\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.433007 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-dns-svc\") pod \"dnsmasq-dns-5b856c5697-9dldl\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.433030 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-9dldl\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.535193 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smlh4\" (UniqueName: \"kubernetes.io/projected/501d57ed-62bc-452a-b0ae-12835a26cd67-kube-api-access-smlh4\") pod \"dnsmasq-dns-5b856c5697-9dldl\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.535332 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-9dldl\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.535388 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-config\") pod \"dnsmasq-dns-5b856c5697-9dldl\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.535452 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-dns-svc\") pod \"dnsmasq-dns-5b856c5697-9dldl\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.535488 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-9dldl\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.536442 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-dns-svc\") pod \"dnsmasq-dns-5b856c5697-9dldl\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.536579 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-config\") pod \"dnsmasq-dns-5b856c5697-9dldl\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.536693 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-ovsdbserver-sb\") pod \"dnsmasq-dns-5b856c5697-9dldl\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.537073 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-ovsdbserver-nb\") pod \"dnsmasq-dns-5b856c5697-9dldl\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.553764 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smlh4\" (UniqueName: \"kubernetes.io/projected/501d57ed-62bc-452a-b0ae-12835a26cd67-kube-api-access-smlh4\") pod \"dnsmasq-dns-5b856c5697-9dldl\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:28 crc kubenswrapper[4739]: I1201 15:55:28.683106 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:29 crc kubenswrapper[4739]: I1201 15:55:29.186759 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-9dldl"] Dec 01 15:55:30 crc kubenswrapper[4739]: I1201 15:55:30.171390 4739 generic.go:334] "Generic (PLEG): container finished" podID="501d57ed-62bc-452a-b0ae-12835a26cd67" containerID="c6e7168e2b20ac0f450cbbe8f5ea9083625ed28eafeb9b4edcf2439928d01057" exitCode=0 Dec 01 15:55:30 crc kubenswrapper[4739]: I1201 15:55:30.171470 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-9dldl" event={"ID":"501d57ed-62bc-452a-b0ae-12835a26cd67","Type":"ContainerDied","Data":"c6e7168e2b20ac0f450cbbe8f5ea9083625ed28eafeb9b4edcf2439928d01057"} Dec 01 15:55:30 crc kubenswrapper[4739]: I1201 15:55:30.172453 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-9dldl" event={"ID":"501d57ed-62bc-452a-b0ae-12835a26cd67","Type":"ContainerStarted","Data":"23c8e234b2b3dbbe00cff6e4b7e02ebee4f5c2410cf93d1db208b36bd0ec0838"} Dec 01 15:55:30 crc kubenswrapper[4739]: I1201 15:55:30.793810 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:55:30 crc kubenswrapper[4739]: I1201 15:55:30.795630 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="proxy-httpd" containerID="cri-o://2fce24f366ea490a132aca38413502b104477bb84c102c3bab3b06fc5b645398" gracePeriod=30 Dec 01 15:55:30 crc kubenswrapper[4739]: I1201 15:55:30.795656 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="sg-core" containerID="cri-o://fedb362e102f965cec0fb488edbee589d98f7594016a736e34fe4f0c29698e42" gracePeriod=30 Dec 01 15:55:30 crc kubenswrapper[4739]: I1201 15:55:30.795683 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="ceilometer-notification-agent" containerID="cri-o://eef22ca6c668d6f576f86a6e3d69bc496172a0250fcf9a88cc4dbe3962bd46f6" gracePeriod=30 Dec 01 15:55:30 crc kubenswrapper[4739]: I1201 15:55:30.796055 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="ceilometer-central-agent" containerID="cri-o://d660793640981a88857238007f7dc9dbc3b55c7fe2d3a50eaeacb8d5829c8555" gracePeriod=30 Dec 01 15:55:31 crc kubenswrapper[4739]: I1201 15:55:31.161297 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:31 crc kubenswrapper[4739]: I1201 15:55:31.187504 4739 generic.go:334] "Generic (PLEG): container finished" podID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerID="2fce24f366ea490a132aca38413502b104477bb84c102c3bab3b06fc5b645398" exitCode=0 Dec 01 15:55:31 crc kubenswrapper[4739]: I1201 15:55:31.187543 4739 generic.go:334] "Generic (PLEG): container finished" podID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerID="fedb362e102f965cec0fb488edbee589d98f7594016a736e34fe4f0c29698e42" exitCode=2 Dec 01 15:55:31 crc kubenswrapper[4739]: I1201 15:55:31.187557 4739 generic.go:334] "Generic (PLEG): container finished" podID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerID="d660793640981a88857238007f7dc9dbc3b55c7fe2d3a50eaeacb8d5829c8555" exitCode=0 Dec 01 15:55:31 crc kubenswrapper[4739]: I1201 15:55:31.187600 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb4aa436-d5da-4b15-b0d5-dcad7118be13","Type":"ContainerDied","Data":"2fce24f366ea490a132aca38413502b104477bb84c102c3bab3b06fc5b645398"} Dec 01 15:55:31 crc kubenswrapper[4739]: I1201 15:55:31.187657 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb4aa436-d5da-4b15-b0d5-dcad7118be13","Type":"ContainerDied","Data":"fedb362e102f965cec0fb488edbee589d98f7594016a736e34fe4f0c29698e42"} Dec 01 15:55:31 crc kubenswrapper[4739]: I1201 15:55:31.187673 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb4aa436-d5da-4b15-b0d5-dcad7118be13","Type":"ContainerDied","Data":"d660793640981a88857238007f7dc9dbc3b55c7fe2d3a50eaeacb8d5829c8555"} Dec 01 15:55:31 crc kubenswrapper[4739]: I1201 15:55:31.190617 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-9dldl" event={"ID":"501d57ed-62bc-452a-b0ae-12835a26cd67","Type":"ContainerStarted","Data":"de16423d8cc93ba215ca89de38a21ace1ff64866de049b16ff82087653c2de36"} Dec 01 15:55:31 crc kubenswrapper[4739]: I1201 15:55:31.190768 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7ed721c3-b669-4601-a166-d851e7960b2c" containerName="nova-api-log" containerID="cri-o://833e1e56711e5b5c55b51087caee82082231fc03077d4f3b52b9c5579c45ccc1" gracePeriod=30 Dec 01 15:55:31 crc kubenswrapper[4739]: I1201 15:55:31.190880 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7ed721c3-b669-4601-a166-d851e7960b2c" containerName="nova-api-api" containerID="cri-o://9ebc3a7b9cabc58ccb10e9b42226d4467672ac309fcc8cd15f09e88838831068" gracePeriod=30 Dec 01 15:55:31 crc kubenswrapper[4739]: I1201 15:55:31.220985 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b856c5697-9dldl" podStartSLOduration=3.220964188 podStartE2EDuration="3.220964188s" podCreationTimestamp="2025-12-01 15:55:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:55:31.215256902 +0000 UTC m=+1233.041003006" watchObservedRunningTime="2025-12-01 15:55:31.220964188 +0000 UTC m=+1233.046710282" Dec 01 15:55:31 crc kubenswrapper[4739]: I1201 15:55:31.582295 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.200699 4739 generic.go:334] "Generic (PLEG): container finished" podID="7ed721c3-b669-4601-a166-d851e7960b2c" containerID="833e1e56711e5b5c55b51087caee82082231fc03077d4f3b52b9c5579c45ccc1" exitCode=143 Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.201040 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ed721c3-b669-4601-a166-d851e7960b2c","Type":"ContainerDied","Data":"833e1e56711e5b5c55b51087caee82082231fc03077d4f3b52b9c5579c45ccc1"} Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.203046 4739 generic.go:334] "Generic (PLEG): container finished" podID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerID="eef22ca6c668d6f576f86a6e3d69bc496172a0250fcf9a88cc4dbe3962bd46f6" exitCode=0 Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.203962 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb4aa436-d5da-4b15-b0d5-dcad7118be13","Type":"ContainerDied","Data":"eef22ca6c668d6f576f86a6e3d69bc496172a0250fcf9a88cc4dbe3962bd46f6"} Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.203994 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.301750 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.436550 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsg88\" (UniqueName: \"kubernetes.io/projected/fb4aa436-d5da-4b15-b0d5-dcad7118be13-kube-api-access-rsg88\") pod \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.436647 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb4aa436-d5da-4b15-b0d5-dcad7118be13-log-httpd\") pod \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.436692 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-ceilometer-tls-certs\") pod \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.436751 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-sg-core-conf-yaml\") pod \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.436794 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb4aa436-d5da-4b15-b0d5-dcad7118be13-run-httpd\") pod \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.436831 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-config-data\") pod \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.436893 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-combined-ca-bundle\") pod \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.436921 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-scripts\") pod \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\" (UID: \"fb4aa436-d5da-4b15-b0d5-dcad7118be13\") " Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.438745 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb4aa436-d5da-4b15-b0d5-dcad7118be13-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fb4aa436-d5da-4b15-b0d5-dcad7118be13" (UID: "fb4aa436-d5da-4b15-b0d5-dcad7118be13"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.439272 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb4aa436-d5da-4b15-b0d5-dcad7118be13-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fb4aa436-d5da-4b15-b0d5-dcad7118be13" (UID: "fb4aa436-d5da-4b15-b0d5-dcad7118be13"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.447241 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-scripts" (OuterVolumeSpecName: "scripts") pod "fb4aa436-d5da-4b15-b0d5-dcad7118be13" (UID: "fb4aa436-d5da-4b15-b0d5-dcad7118be13"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.447258 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb4aa436-d5da-4b15-b0d5-dcad7118be13-kube-api-access-rsg88" (OuterVolumeSpecName: "kube-api-access-rsg88") pod "fb4aa436-d5da-4b15-b0d5-dcad7118be13" (UID: "fb4aa436-d5da-4b15-b0d5-dcad7118be13"). InnerVolumeSpecName "kube-api-access-rsg88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.491657 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fb4aa436-d5da-4b15-b0d5-dcad7118be13" (UID: "fb4aa436-d5da-4b15-b0d5-dcad7118be13"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.519097 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "fb4aa436-d5da-4b15-b0d5-dcad7118be13" (UID: "fb4aa436-d5da-4b15-b0d5-dcad7118be13"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.528862 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb4aa436-d5da-4b15-b0d5-dcad7118be13" (UID: "fb4aa436-d5da-4b15-b0d5-dcad7118be13"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.539176 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsg88\" (UniqueName: \"kubernetes.io/projected/fb4aa436-d5da-4b15-b0d5-dcad7118be13-kube-api-access-rsg88\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.539223 4739 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb4aa436-d5da-4b15-b0d5-dcad7118be13-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.539241 4739 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.539256 4739 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.539269 4739 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb4aa436-d5da-4b15-b0d5-dcad7118be13-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.539283 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.539297 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.548838 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-config-data" (OuterVolumeSpecName: "config-data") pod "fb4aa436-d5da-4b15-b0d5-dcad7118be13" (UID: "fb4aa436-d5da-4b15-b0d5-dcad7118be13"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:32 crc kubenswrapper[4739]: I1201 15:55:32.640552 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb4aa436-d5da-4b15-b0d5-dcad7118be13-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.218077 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb4aa436-d5da-4b15-b0d5-dcad7118be13","Type":"ContainerDied","Data":"67ac5e1768e476ec7c75f48fce936ef9e56d1829281e8f40031fafda107a4b4b"} Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.218185 4739 scope.go:117] "RemoveContainer" containerID="2fce24f366ea490a132aca38413502b104477bb84c102c3bab3b06fc5b645398" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.218216 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.247079 4739 scope.go:117] "RemoveContainer" containerID="fedb362e102f965cec0fb488edbee589d98f7594016a736e34fe4f0c29698e42" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.270464 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.275355 4739 scope.go:117] "RemoveContainer" containerID="eef22ca6c668d6f576f86a6e3d69bc496172a0250fcf9a88cc4dbe3962bd46f6" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.282613 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.298044 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:55:33 crc kubenswrapper[4739]: E1201 15:55:33.298932 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="ceilometer-notification-agent" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.298954 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="ceilometer-notification-agent" Dec 01 15:55:33 crc kubenswrapper[4739]: E1201 15:55:33.298971 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="ceilometer-central-agent" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.298996 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="ceilometer-central-agent" Dec 01 15:55:33 crc kubenswrapper[4739]: E1201 15:55:33.299017 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="sg-core" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.299025 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="sg-core" Dec 01 15:55:33 crc kubenswrapper[4739]: E1201 15:55:33.299056 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="proxy-httpd" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.299062 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="proxy-httpd" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.299218 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="sg-core" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.299238 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="ceilometer-central-agent" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.299251 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="proxy-httpd" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.299259 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" containerName="ceilometer-notification-agent" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.301147 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.301919 4739 scope.go:117] "RemoveContainer" containerID="d660793640981a88857238007f7dc9dbc3b55c7fe2d3a50eaeacb8d5829c8555" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.304126 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.304358 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.304495 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.314167 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.455176 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.455250 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37c0dad8-d954-4bae-8654-68085a9eb6d3-log-httpd\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.455407 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-config-data\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.455655 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-scripts\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.455685 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk4h4\" (UniqueName: \"kubernetes.io/projected/37c0dad8-d954-4bae-8654-68085a9eb6d3-kube-api-access-dk4h4\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.455709 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37c0dad8-d954-4bae-8654-68085a9eb6d3-run-httpd\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.455943 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.455992 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.557571 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.557693 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37c0dad8-d954-4bae-8654-68085a9eb6d3-log-httpd\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.557743 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-config-data\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.557840 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-scripts\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.557875 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37c0dad8-d954-4bae-8654-68085a9eb6d3-run-httpd\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.557905 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk4h4\" (UniqueName: \"kubernetes.io/projected/37c0dad8-d954-4bae-8654-68085a9eb6d3-kube-api-access-dk4h4\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.558066 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.558099 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.559108 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37c0dad8-d954-4bae-8654-68085a9eb6d3-run-httpd\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.559469 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37c0dad8-d954-4bae-8654-68085a9eb6d3-log-httpd\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.565888 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.567235 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.570145 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.585832 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-config-data\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.587588 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-scripts\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.596651 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk4h4\" (UniqueName: \"kubernetes.io/projected/37c0dad8-d954-4bae-8654-68085a9eb6d3-kube-api-access-dk4h4\") pod \"ceilometer-0\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " pod="openstack/ceilometer-0" Dec 01 15:55:33 crc kubenswrapper[4739]: I1201 15:55:33.632019 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.102151 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 15:55:34 crc kubenswrapper[4739]: W1201 15:55:34.115725 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37c0dad8_d954_4bae_8654_68085a9eb6d3.slice/crio-345c8dbccec64e8b4bf7ed6c15474b8ebf928d6ef8a6d2ec0cdefe8c88ff3e65 WatchSource:0}: Error finding container 345c8dbccec64e8b4bf7ed6c15474b8ebf928d6ef8a6d2ec0cdefe8c88ff3e65: Status 404 returned error can't find the container with id 345c8dbccec64e8b4bf7ed6c15474b8ebf928d6ef8a6d2ec0cdefe8c88ff3e65 Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.118715 4739 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.228938 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37c0dad8-d954-4bae-8654-68085a9eb6d3","Type":"ContainerStarted","Data":"345c8dbccec64e8b4bf7ed6c15474b8ebf928d6ef8a6d2ec0cdefe8c88ff3e65"} Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.493057 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb4aa436-d5da-4b15-b0d5-dcad7118be13" path="/var/lib/kubelet/pods/fb4aa436-d5da-4b15-b0d5-dcad7118be13/volumes" Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.711664 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.882212 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ed721c3-b669-4601-a166-d851e7960b2c-config-data\") pod \"7ed721c3-b669-4601-a166-d851e7960b2c\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.882279 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f48fl\" (UniqueName: \"kubernetes.io/projected/7ed721c3-b669-4601-a166-d851e7960b2c-kube-api-access-f48fl\") pod \"7ed721c3-b669-4601-a166-d851e7960b2c\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.882560 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ed721c3-b669-4601-a166-d851e7960b2c-combined-ca-bundle\") pod \"7ed721c3-b669-4601-a166-d851e7960b2c\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.882645 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ed721c3-b669-4601-a166-d851e7960b2c-logs\") pod \"7ed721c3-b669-4601-a166-d851e7960b2c\" (UID: \"7ed721c3-b669-4601-a166-d851e7960b2c\") " Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.883323 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ed721c3-b669-4601-a166-d851e7960b2c-logs" (OuterVolumeSpecName: "logs") pod "7ed721c3-b669-4601-a166-d851e7960b2c" (UID: "7ed721c3-b669-4601-a166-d851e7960b2c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.887046 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ed721c3-b669-4601-a166-d851e7960b2c-kube-api-access-f48fl" (OuterVolumeSpecName: "kube-api-access-f48fl") pod "7ed721c3-b669-4601-a166-d851e7960b2c" (UID: "7ed721c3-b669-4601-a166-d851e7960b2c"). InnerVolumeSpecName "kube-api-access-f48fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.908601 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ed721c3-b669-4601-a166-d851e7960b2c-config-data" (OuterVolumeSpecName: "config-data") pod "7ed721c3-b669-4601-a166-d851e7960b2c" (UID: "7ed721c3-b669-4601-a166-d851e7960b2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.922329 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ed721c3-b669-4601-a166-d851e7960b2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ed721c3-b669-4601-a166-d851e7960b2c" (UID: "7ed721c3-b669-4601-a166-d851e7960b2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.984853 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ed721c3-b669-4601-a166-d851e7960b2c-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.984883 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f48fl\" (UniqueName: \"kubernetes.io/projected/7ed721c3-b669-4601-a166-d851e7960b2c-kube-api-access-f48fl\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.984894 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ed721c3-b669-4601-a166-d851e7960b2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:34 crc kubenswrapper[4739]: I1201 15:55:34.984903 4739 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ed721c3-b669-4601-a166-d851e7960b2c-logs\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.242505 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37c0dad8-d954-4bae-8654-68085a9eb6d3","Type":"ContainerStarted","Data":"5afacc5853ae439e6e32109942743089d84b1baeba5b9df62165dbf87ab3c2f4"} Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.245711 4739 generic.go:334] "Generic (PLEG): container finished" podID="7ed721c3-b669-4601-a166-d851e7960b2c" containerID="9ebc3a7b9cabc58ccb10e9b42226d4467672ac309fcc8cd15f09e88838831068" exitCode=0 Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.245774 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.245778 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ed721c3-b669-4601-a166-d851e7960b2c","Type":"ContainerDied","Data":"9ebc3a7b9cabc58ccb10e9b42226d4467672ac309fcc8cd15f09e88838831068"} Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.245965 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ed721c3-b669-4601-a166-d851e7960b2c","Type":"ContainerDied","Data":"7da751400ea2fdde0fab17feff61a00ce2d72d6321b14542df24c9c64e538921"} Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.246035 4739 scope.go:117] "RemoveContainer" containerID="9ebc3a7b9cabc58ccb10e9b42226d4467672ac309fcc8cd15f09e88838831068" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.275748 4739 scope.go:117] "RemoveContainer" containerID="833e1e56711e5b5c55b51087caee82082231fc03077d4f3b52b9c5579c45ccc1" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.309805 4739 scope.go:117] "RemoveContainer" containerID="9ebc3a7b9cabc58ccb10e9b42226d4467672ac309fcc8cd15f09e88838831068" Dec 01 15:55:35 crc kubenswrapper[4739]: E1201 15:55:35.313972 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ebc3a7b9cabc58ccb10e9b42226d4467672ac309fcc8cd15f09e88838831068\": container with ID starting with 9ebc3a7b9cabc58ccb10e9b42226d4467672ac309fcc8cd15f09e88838831068 not found: ID does not exist" containerID="9ebc3a7b9cabc58ccb10e9b42226d4467672ac309fcc8cd15f09e88838831068" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.314012 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ebc3a7b9cabc58ccb10e9b42226d4467672ac309fcc8cd15f09e88838831068"} err="failed to get container status \"9ebc3a7b9cabc58ccb10e9b42226d4467672ac309fcc8cd15f09e88838831068\": rpc error: code = NotFound desc = could not find container \"9ebc3a7b9cabc58ccb10e9b42226d4467672ac309fcc8cd15f09e88838831068\": container with ID starting with 9ebc3a7b9cabc58ccb10e9b42226d4467672ac309fcc8cd15f09e88838831068 not found: ID does not exist" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.314038 4739 scope.go:117] "RemoveContainer" containerID="833e1e56711e5b5c55b51087caee82082231fc03077d4f3b52b9c5579c45ccc1" Dec 01 15:55:35 crc kubenswrapper[4739]: E1201 15:55:35.319977 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"833e1e56711e5b5c55b51087caee82082231fc03077d4f3b52b9c5579c45ccc1\": container with ID starting with 833e1e56711e5b5c55b51087caee82082231fc03077d4f3b52b9c5579c45ccc1 not found: ID does not exist" containerID="833e1e56711e5b5c55b51087caee82082231fc03077d4f3b52b9c5579c45ccc1" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.320011 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"833e1e56711e5b5c55b51087caee82082231fc03077d4f3b52b9c5579c45ccc1"} err="failed to get container status \"833e1e56711e5b5c55b51087caee82082231fc03077d4f3b52b9c5579c45ccc1\": rpc error: code = NotFound desc = could not find container \"833e1e56711e5b5c55b51087caee82082231fc03077d4f3b52b9c5579c45ccc1\": container with ID starting with 833e1e56711e5b5c55b51087caee82082231fc03077d4f3b52b9c5579c45ccc1 not found: ID does not exist" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.342533 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.372755 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.386567 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:35 crc kubenswrapper[4739]: E1201 15:55:35.387133 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ed721c3-b669-4601-a166-d851e7960b2c" containerName="nova-api-api" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.387161 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ed721c3-b669-4601-a166-d851e7960b2c" containerName="nova-api-api" Dec 01 15:55:35 crc kubenswrapper[4739]: E1201 15:55:35.387198 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ed721c3-b669-4601-a166-d851e7960b2c" containerName="nova-api-log" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.387206 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ed721c3-b669-4601-a166-d851e7960b2c" containerName="nova-api-log" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.387684 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ed721c3-b669-4601-a166-d851e7960b2c" containerName="nova-api-api" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.387758 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ed721c3-b669-4601-a166-d851e7960b2c" containerName="nova-api-log" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.389064 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.390999 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.391270 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.391398 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.393761 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.506644 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-config-data\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.506727 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.506867 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.506926 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fdac5301-3c64-40a8-85b3-67a92b22fa2a-logs\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.506983 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5865f\" (UniqueName: \"kubernetes.io/projected/fdac5301-3c64-40a8-85b3-67a92b22fa2a-kube-api-access-5865f\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.507051 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-public-tls-certs\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.609339 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.609463 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fdac5301-3c64-40a8-85b3-67a92b22fa2a-logs\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.609522 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5865f\" (UniqueName: \"kubernetes.io/projected/fdac5301-3c64-40a8-85b3-67a92b22fa2a-kube-api-access-5865f\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.609596 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-public-tls-certs\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.609674 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-config-data\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.609731 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.610991 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fdac5301-3c64-40a8-85b3-67a92b22fa2a-logs\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.613932 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-public-tls-certs\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.614072 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.615079 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.615453 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-config-data\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.631871 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5865f\" (UniqueName: \"kubernetes.io/projected/fdac5301-3c64-40a8-85b3-67a92b22fa2a-kube-api-access-5865f\") pod \"nova-api-0\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " pod="openstack/nova-api-0" Dec 01 15:55:35 crc kubenswrapper[4739]: I1201 15:55:35.717090 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 15:55:36 crc kubenswrapper[4739]: I1201 15:55:36.150966 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:36 crc kubenswrapper[4739]: W1201 15:55:36.151645 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdac5301_3c64_40a8_85b3_67a92b22fa2a.slice/crio-08f85c78104f95579e599ab854ce71fbc40114c60c0d77ddd8d30a6e1d887357 WatchSource:0}: Error finding container 08f85c78104f95579e599ab854ce71fbc40114c60c0d77ddd8d30a6e1d887357: Status 404 returned error can't find the container with id 08f85c78104f95579e599ab854ce71fbc40114c60c0d77ddd8d30a6e1d887357 Dec 01 15:55:36 crc kubenswrapper[4739]: I1201 15:55:36.263482 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37c0dad8-d954-4bae-8654-68085a9eb6d3","Type":"ContainerStarted","Data":"7cea978622b07b42b7ec40beeec9a14232afcc1ad2b4e521453f7d8fb9d8eb12"} Dec 01 15:55:36 crc kubenswrapper[4739]: I1201 15:55:36.267407 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fdac5301-3c64-40a8-85b3-67a92b22fa2a","Type":"ContainerStarted","Data":"08f85c78104f95579e599ab854ce71fbc40114c60c0d77ddd8d30a6e1d887357"} Dec 01 15:55:36 crc kubenswrapper[4739]: I1201 15:55:36.489675 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ed721c3-b669-4601-a166-d851e7960b2c" path="/var/lib/kubelet/pods/7ed721c3-b669-4601-a166-d851e7960b2c/volumes" Dec 01 15:55:36 crc kubenswrapper[4739]: I1201 15:55:36.582662 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:36 crc kubenswrapper[4739]: I1201 15:55:36.601579 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.285321 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37c0dad8-d954-4bae-8654-68085a9eb6d3","Type":"ContainerStarted","Data":"5954d26bc5a08aa1686874d52e97a553d0c0a97f6083ae9eda383c70da6683f6"} Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.287496 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fdac5301-3c64-40a8-85b3-67a92b22fa2a","Type":"ContainerStarted","Data":"fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98"} Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.287517 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fdac5301-3c64-40a8-85b3-67a92b22fa2a","Type":"ContainerStarted","Data":"0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640"} Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.303633 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.314727 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.314707765 podStartE2EDuration="2.314707765s" podCreationTimestamp="2025-12-01 15:55:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:55:37.310130903 +0000 UTC m=+1239.135877037" watchObservedRunningTime="2025-12-01 15:55:37.314707765 +0000 UTC m=+1239.140453869" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.505499 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-sh566"] Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.506722 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.508858 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.516729 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.535759 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-sh566"] Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.646370 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-config-data\") pod \"nova-cell1-cell-mapping-sh566\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.646454 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8hzn\" (UniqueName: \"kubernetes.io/projected/951ace80-f656-4e64-85e7-d4d54d793159-kube-api-access-f8hzn\") pod \"nova-cell1-cell-mapping-sh566\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.646549 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-scripts\") pod \"nova-cell1-cell-mapping-sh566\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.646640 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-sh566\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.748460 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-config-data\") pod \"nova-cell1-cell-mapping-sh566\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.748515 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8hzn\" (UniqueName: \"kubernetes.io/projected/951ace80-f656-4e64-85e7-d4d54d793159-kube-api-access-f8hzn\") pod \"nova-cell1-cell-mapping-sh566\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.748578 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-scripts\") pod \"nova-cell1-cell-mapping-sh566\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.748663 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-sh566\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.753582 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-config-data\") pod \"nova-cell1-cell-mapping-sh566\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.754107 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-scripts\") pod \"nova-cell1-cell-mapping-sh566\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.757975 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-sh566\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.769488 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8hzn\" (UniqueName: \"kubernetes.io/projected/951ace80-f656-4e64-85e7-d4d54d793159-kube-api-access-f8hzn\") pod \"nova-cell1-cell-mapping-sh566\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:37 crc kubenswrapper[4739]: I1201 15:55:37.830901 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:38 crc kubenswrapper[4739]: I1201 15:55:38.302898 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37c0dad8-d954-4bae-8654-68085a9eb6d3","Type":"ContainerStarted","Data":"710683be135103ccee71ba86e78c898cd9863ecfe958ba0b6783e508b9b83998"} Dec 01 15:55:38 crc kubenswrapper[4739]: I1201 15:55:38.310231 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-sh566"] Dec 01 15:55:38 crc kubenswrapper[4739]: I1201 15:55:38.345831 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.6025146559999999 podStartE2EDuration="5.345811566s" podCreationTimestamp="2025-12-01 15:55:33 +0000 UTC" firstStartedPulling="2025-12-01 15:55:34.118401224 +0000 UTC m=+1235.944147318" lastFinishedPulling="2025-12-01 15:55:37.861698134 +0000 UTC m=+1239.687444228" observedRunningTime="2025-12-01 15:55:38.342934556 +0000 UTC m=+1240.168680650" watchObservedRunningTime="2025-12-01 15:55:38.345811566 +0000 UTC m=+1240.171557680" Dec 01 15:55:38 crc kubenswrapper[4739]: I1201 15:55:38.684626 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:55:38 crc kubenswrapper[4739]: I1201 15:55:38.765237 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-x89dh"] Dec 01 15:55:38 crc kubenswrapper[4739]: I1201 15:55:38.765682 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-566b5b7845-x89dh" podUID="eb53ff5e-3cee-4686-befd-d071117f4541" containerName="dnsmasq-dns" containerID="cri-o://08302d554bc32eb4521c43e320f19d3a45fbac9b4964ec745918cfee15b5965a" gracePeriod=10 Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.277557 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.317778 4739 generic.go:334] "Generic (PLEG): container finished" podID="eb53ff5e-3cee-4686-befd-d071117f4541" containerID="08302d554bc32eb4521c43e320f19d3a45fbac9b4964ec745918cfee15b5965a" exitCode=0 Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.317852 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-x89dh" event={"ID":"eb53ff5e-3cee-4686-befd-d071117f4541","Type":"ContainerDied","Data":"08302d554bc32eb4521c43e320f19d3a45fbac9b4964ec745918cfee15b5965a"} Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.317885 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-566b5b7845-x89dh" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.318672 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-566b5b7845-x89dh" event={"ID":"eb53ff5e-3cee-4686-befd-d071117f4541","Type":"ContainerDied","Data":"09e5f232ae637ff01321a16712f4dc7ddd20fc1dbbc6d4b5f2b67e1dc40c826d"} Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.318709 4739 scope.go:117] "RemoveContainer" containerID="08302d554bc32eb4521c43e320f19d3a45fbac9b4964ec745918cfee15b5965a" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.322535 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-sh566" event={"ID":"951ace80-f656-4e64-85e7-d4d54d793159","Type":"ContainerStarted","Data":"9bf623b03225199969eac35676c02ac9a56eddd549610375275b661f7308174e"} Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.322579 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.322594 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-sh566" event={"ID":"951ace80-f656-4e64-85e7-d4d54d793159","Type":"ContainerStarted","Data":"7a9085147130be7afd269c5294dfbfc1ec918a37a8b0e0d61955d27f0ec27713"} Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.342876 4739 scope.go:117] "RemoveContainer" containerID="fdb7a80be083b941db94ed60ecef17fc6a9bad38033f2342ba4202e43525f688" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.346592 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-sh566" podStartSLOduration=2.346574876 podStartE2EDuration="2.346574876s" podCreationTimestamp="2025-12-01 15:55:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:55:39.343284224 +0000 UTC m=+1241.169030328" watchObservedRunningTime="2025-12-01 15:55:39.346574876 +0000 UTC m=+1241.172320970" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.378685 4739 scope.go:117] "RemoveContainer" containerID="08302d554bc32eb4521c43e320f19d3a45fbac9b4964ec745918cfee15b5965a" Dec 01 15:55:39 crc kubenswrapper[4739]: E1201 15:55:39.379122 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08302d554bc32eb4521c43e320f19d3a45fbac9b4964ec745918cfee15b5965a\": container with ID starting with 08302d554bc32eb4521c43e320f19d3a45fbac9b4964ec745918cfee15b5965a not found: ID does not exist" containerID="08302d554bc32eb4521c43e320f19d3a45fbac9b4964ec745918cfee15b5965a" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.379165 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08302d554bc32eb4521c43e320f19d3a45fbac9b4964ec745918cfee15b5965a"} err="failed to get container status \"08302d554bc32eb4521c43e320f19d3a45fbac9b4964ec745918cfee15b5965a\": rpc error: code = NotFound desc = could not find container \"08302d554bc32eb4521c43e320f19d3a45fbac9b4964ec745918cfee15b5965a\": container with ID starting with 08302d554bc32eb4521c43e320f19d3a45fbac9b4964ec745918cfee15b5965a not found: ID does not exist" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.379191 4739 scope.go:117] "RemoveContainer" containerID="fdb7a80be083b941db94ed60ecef17fc6a9bad38033f2342ba4202e43525f688" Dec 01 15:55:39 crc kubenswrapper[4739]: E1201 15:55:39.379646 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdb7a80be083b941db94ed60ecef17fc6a9bad38033f2342ba4202e43525f688\": container with ID starting with fdb7a80be083b941db94ed60ecef17fc6a9bad38033f2342ba4202e43525f688 not found: ID does not exist" containerID="fdb7a80be083b941db94ed60ecef17fc6a9bad38033f2342ba4202e43525f688" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.379676 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdb7a80be083b941db94ed60ecef17fc6a9bad38033f2342ba4202e43525f688"} err="failed to get container status \"fdb7a80be083b941db94ed60ecef17fc6a9bad38033f2342ba4202e43525f688\": rpc error: code = NotFound desc = could not find container \"fdb7a80be083b941db94ed60ecef17fc6a9bad38033f2342ba4202e43525f688\": container with ID starting with fdb7a80be083b941db94ed60ecef17fc6a9bad38033f2342ba4202e43525f688 not found: ID does not exist" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.382097 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwfp4\" (UniqueName: \"kubernetes.io/projected/eb53ff5e-3cee-4686-befd-d071117f4541-kube-api-access-dwfp4\") pod \"eb53ff5e-3cee-4686-befd-d071117f4541\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.382146 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-ovsdbserver-nb\") pod \"eb53ff5e-3cee-4686-befd-d071117f4541\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.382196 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-config\") pod \"eb53ff5e-3cee-4686-befd-d071117f4541\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.382268 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-ovsdbserver-sb\") pod \"eb53ff5e-3cee-4686-befd-d071117f4541\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.382317 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-dns-svc\") pod \"eb53ff5e-3cee-4686-befd-d071117f4541\" (UID: \"eb53ff5e-3cee-4686-befd-d071117f4541\") " Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.387692 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb53ff5e-3cee-4686-befd-d071117f4541-kube-api-access-dwfp4" (OuterVolumeSpecName: "kube-api-access-dwfp4") pod "eb53ff5e-3cee-4686-befd-d071117f4541" (UID: "eb53ff5e-3cee-4686-befd-d071117f4541"). InnerVolumeSpecName "kube-api-access-dwfp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.423930 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eb53ff5e-3cee-4686-befd-d071117f4541" (UID: "eb53ff5e-3cee-4686-befd-d071117f4541"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.429317 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eb53ff5e-3cee-4686-befd-d071117f4541" (UID: "eb53ff5e-3cee-4686-befd-d071117f4541"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.431784 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eb53ff5e-3cee-4686-befd-d071117f4541" (UID: "eb53ff5e-3cee-4686-befd-d071117f4541"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.441952 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-config" (OuterVolumeSpecName: "config") pod "eb53ff5e-3cee-4686-befd-d071117f4541" (UID: "eb53ff5e-3cee-4686-befd-d071117f4541"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.484228 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwfp4\" (UniqueName: \"kubernetes.io/projected/eb53ff5e-3cee-4686-befd-d071117f4541-kube-api-access-dwfp4\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.484258 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.484268 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.484280 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.484293 4739 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb53ff5e-3cee-4686-befd-d071117f4541-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.621491 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.621562 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.653808 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-x89dh"] Dec 01 15:55:39 crc kubenswrapper[4739]: I1201 15:55:39.662651 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-566b5b7845-x89dh"] Dec 01 15:55:40 crc kubenswrapper[4739]: I1201 15:55:40.495610 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb53ff5e-3cee-4686-befd-d071117f4541" path="/var/lib/kubelet/pods/eb53ff5e-3cee-4686-befd-d071117f4541/volumes" Dec 01 15:55:43 crc kubenswrapper[4739]: I1201 15:55:43.368033 4739 generic.go:334] "Generic (PLEG): container finished" podID="951ace80-f656-4e64-85e7-d4d54d793159" containerID="9bf623b03225199969eac35676c02ac9a56eddd549610375275b661f7308174e" exitCode=0 Dec 01 15:55:43 crc kubenswrapper[4739]: I1201 15:55:43.368083 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-sh566" event={"ID":"951ace80-f656-4e64-85e7-d4d54d793159","Type":"ContainerDied","Data":"9bf623b03225199969eac35676c02ac9a56eddd549610375275b661f7308174e"} Dec 01 15:55:44 crc kubenswrapper[4739]: I1201 15:55:44.796879 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:44 crc kubenswrapper[4739]: I1201 15:55:44.908535 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-combined-ca-bundle\") pod \"951ace80-f656-4e64-85e7-d4d54d793159\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " Dec 01 15:55:44 crc kubenswrapper[4739]: I1201 15:55:44.908643 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-scripts\") pod \"951ace80-f656-4e64-85e7-d4d54d793159\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " Dec 01 15:55:44 crc kubenswrapper[4739]: I1201 15:55:44.908862 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-config-data\") pod \"951ace80-f656-4e64-85e7-d4d54d793159\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " Dec 01 15:55:44 crc kubenswrapper[4739]: I1201 15:55:44.908947 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8hzn\" (UniqueName: \"kubernetes.io/projected/951ace80-f656-4e64-85e7-d4d54d793159-kube-api-access-f8hzn\") pod \"951ace80-f656-4e64-85e7-d4d54d793159\" (UID: \"951ace80-f656-4e64-85e7-d4d54d793159\") " Dec 01 15:55:44 crc kubenswrapper[4739]: I1201 15:55:44.919021 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-scripts" (OuterVolumeSpecName: "scripts") pod "951ace80-f656-4e64-85e7-d4d54d793159" (UID: "951ace80-f656-4e64-85e7-d4d54d793159"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:44 crc kubenswrapper[4739]: I1201 15:55:44.919040 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/951ace80-f656-4e64-85e7-d4d54d793159-kube-api-access-f8hzn" (OuterVolumeSpecName: "kube-api-access-f8hzn") pod "951ace80-f656-4e64-85e7-d4d54d793159" (UID: "951ace80-f656-4e64-85e7-d4d54d793159"). InnerVolumeSpecName "kube-api-access-f8hzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:44 crc kubenswrapper[4739]: I1201 15:55:44.961642 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "951ace80-f656-4e64-85e7-d4d54d793159" (UID: "951ace80-f656-4e64-85e7-d4d54d793159"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:44 crc kubenswrapper[4739]: I1201 15:55:44.962770 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-config-data" (OuterVolumeSpecName: "config-data") pod "951ace80-f656-4e64-85e7-d4d54d793159" (UID: "951ace80-f656-4e64-85e7-d4d54d793159"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:45 crc kubenswrapper[4739]: I1201 15:55:45.011558 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:45 crc kubenswrapper[4739]: I1201 15:55:45.011610 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8hzn\" (UniqueName: \"kubernetes.io/projected/951ace80-f656-4e64-85e7-d4d54d793159-kube-api-access-f8hzn\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:45 crc kubenswrapper[4739]: I1201 15:55:45.011632 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:45 crc kubenswrapper[4739]: I1201 15:55:45.011651 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/951ace80-f656-4e64-85e7-d4d54d793159-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:45 crc kubenswrapper[4739]: I1201 15:55:45.394977 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-sh566" event={"ID":"951ace80-f656-4e64-85e7-d4d54d793159","Type":"ContainerDied","Data":"7a9085147130be7afd269c5294dfbfc1ec918a37a8b0e0d61955d27f0ec27713"} Dec 01 15:55:45 crc kubenswrapper[4739]: I1201 15:55:45.395321 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a9085147130be7afd269c5294dfbfc1ec918a37a8b0e0d61955d27f0ec27713" Dec 01 15:55:45 crc kubenswrapper[4739]: I1201 15:55:45.395556 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-sh566" Dec 01 15:55:45 crc kubenswrapper[4739]: I1201 15:55:45.518234 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 15:55:45 crc kubenswrapper[4739]: I1201 15:55:45.518505 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7a928221-a9c0-4602-81d2-f0dede25dafa" containerName="nova-scheduler-scheduler" containerID="cri-o://ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d" gracePeriod=30 Dec 01 15:55:45 crc kubenswrapper[4739]: I1201 15:55:45.529185 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:45 crc kubenswrapper[4739]: I1201 15:55:45.530444 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fdac5301-3c64-40a8-85b3-67a92b22fa2a" containerName="nova-api-api" containerID="cri-o://fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98" gracePeriod=30 Dec 01 15:55:45 crc kubenswrapper[4739]: I1201 15:55:45.529545 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fdac5301-3c64-40a8-85b3-67a92b22fa2a" containerName="nova-api-log" containerID="cri-o://0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640" gracePeriod=30 Dec 01 15:55:45 crc kubenswrapper[4739]: I1201 15:55:45.586725 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:55:45 crc kubenswrapper[4739]: I1201 15:55:45.587161 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b30c462a-2b55-4187-931e-12e78990741a" containerName="nova-metadata-log" containerID="cri-o://7c67238b1af7b9e356f869c4c76a43339bb2f33a8d2452ecd5d77211e6a33247" gracePeriod=30 Dec 01 15:55:45 crc kubenswrapper[4739]: I1201 15:55:45.587301 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b30c462a-2b55-4187-931e-12e78990741a" containerName="nova-metadata-metadata" containerID="cri-o://6f0e116dd42b56eca373c2ab2adb6c497716ffc059119fd760c3616cb644f343" gracePeriod=30 Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.067249 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.131113 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-internal-tls-certs\") pod \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.131181 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-combined-ca-bundle\") pod \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.131259 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-public-tls-certs\") pod \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.131312 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5865f\" (UniqueName: \"kubernetes.io/projected/fdac5301-3c64-40a8-85b3-67a92b22fa2a-kube-api-access-5865f\") pod \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.131342 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-config-data\") pod \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.131470 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fdac5301-3c64-40a8-85b3-67a92b22fa2a-logs\") pod \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\" (UID: \"fdac5301-3c64-40a8-85b3-67a92b22fa2a\") " Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.131966 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdac5301-3c64-40a8-85b3-67a92b22fa2a-logs" (OuterVolumeSpecName: "logs") pod "fdac5301-3c64-40a8-85b3-67a92b22fa2a" (UID: "fdac5301-3c64-40a8-85b3-67a92b22fa2a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.136758 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdac5301-3c64-40a8-85b3-67a92b22fa2a-kube-api-access-5865f" (OuterVolumeSpecName: "kube-api-access-5865f") pod "fdac5301-3c64-40a8-85b3-67a92b22fa2a" (UID: "fdac5301-3c64-40a8-85b3-67a92b22fa2a"). InnerVolumeSpecName "kube-api-access-5865f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.167002 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-config-data" (OuterVolumeSpecName: "config-data") pod "fdac5301-3c64-40a8-85b3-67a92b22fa2a" (UID: "fdac5301-3c64-40a8-85b3-67a92b22fa2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.179968 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fdac5301-3c64-40a8-85b3-67a92b22fa2a" (UID: "fdac5301-3c64-40a8-85b3-67a92b22fa2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.188480 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "fdac5301-3c64-40a8-85b3-67a92b22fa2a" (UID: "fdac5301-3c64-40a8-85b3-67a92b22fa2a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.203621 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "fdac5301-3c64-40a8-85b3-67a92b22fa2a" (UID: "fdac5301-3c64-40a8-85b3-67a92b22fa2a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.234658 4739 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fdac5301-3c64-40a8-85b3-67a92b22fa2a-logs\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.234705 4739 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.234723 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.234735 4739 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.234774 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5865f\" (UniqueName: \"kubernetes.io/projected/fdac5301-3c64-40a8-85b3-67a92b22fa2a-kube-api-access-5865f\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.234789 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdac5301-3c64-40a8-85b3-67a92b22fa2a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:46 crc kubenswrapper[4739]: E1201 15:55:46.264533 4739 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 15:55:46 crc kubenswrapper[4739]: E1201 15:55:46.265949 4739 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 15:55:46 crc kubenswrapper[4739]: E1201 15:55:46.267415 4739 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 15:55:46 crc kubenswrapper[4739]: E1201 15:55:46.267470 4739 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7a928221-a9c0-4602-81d2-f0dede25dafa" containerName="nova-scheduler-scheduler" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.405542 4739 generic.go:334] "Generic (PLEG): container finished" podID="b30c462a-2b55-4187-931e-12e78990741a" containerID="7c67238b1af7b9e356f869c4c76a43339bb2f33a8d2452ecd5d77211e6a33247" exitCode=143 Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.405617 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b30c462a-2b55-4187-931e-12e78990741a","Type":"ContainerDied","Data":"7c67238b1af7b9e356f869c4c76a43339bb2f33a8d2452ecd5d77211e6a33247"} Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.407586 4739 generic.go:334] "Generic (PLEG): container finished" podID="fdac5301-3c64-40a8-85b3-67a92b22fa2a" containerID="fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98" exitCode=0 Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.407615 4739 generic.go:334] "Generic (PLEG): container finished" podID="fdac5301-3c64-40a8-85b3-67a92b22fa2a" containerID="0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640" exitCode=143 Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.407637 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fdac5301-3c64-40a8-85b3-67a92b22fa2a","Type":"ContainerDied","Data":"fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98"} Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.407676 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.407695 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fdac5301-3c64-40a8-85b3-67a92b22fa2a","Type":"ContainerDied","Data":"0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640"} Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.407718 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fdac5301-3c64-40a8-85b3-67a92b22fa2a","Type":"ContainerDied","Data":"08f85c78104f95579e599ab854ce71fbc40114c60c0d77ddd8d30a6e1d887357"} Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.407746 4739 scope.go:117] "RemoveContainer" containerID="fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.429862 4739 scope.go:117] "RemoveContainer" containerID="0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.482231 4739 scope.go:117] "RemoveContainer" containerID="fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98" Dec 01 15:55:46 crc kubenswrapper[4739]: E1201 15:55:46.482910 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98\": container with ID starting with fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98 not found: ID does not exist" containerID="fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.482936 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98"} err="failed to get container status \"fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98\": rpc error: code = NotFound desc = could not find container \"fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98\": container with ID starting with fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98 not found: ID does not exist" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.482954 4739 scope.go:117] "RemoveContainer" containerID="0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640" Dec 01 15:55:46 crc kubenswrapper[4739]: E1201 15:55:46.483265 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640\": container with ID starting with 0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640 not found: ID does not exist" containerID="0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.483317 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640"} err="failed to get container status \"0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640\": rpc error: code = NotFound desc = could not find container \"0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640\": container with ID starting with 0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640 not found: ID does not exist" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.483347 4739 scope.go:117] "RemoveContainer" containerID="fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.485389 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98"} err="failed to get container status \"fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98\": rpc error: code = NotFound desc = could not find container \"fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98\": container with ID starting with fe70cbe7206a1fe67313f398116a4dc94c05cdc13805edb3fd985a377ef27c98 not found: ID does not exist" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.485431 4739 scope.go:117] "RemoveContainer" containerID="0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.486969 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640"} err="failed to get container status \"0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640\": rpc error: code = NotFound desc = could not find container \"0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640\": container with ID starting with 0d962a68fe18a5a4d1650679ea1754e9d117c1f3a47a6f232a45b21980a2a640 not found: ID does not exist" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.489228 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.489262 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.499483 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:46 crc kubenswrapper[4739]: E1201 15:55:46.500004 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="951ace80-f656-4e64-85e7-d4d54d793159" containerName="nova-manage" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.500023 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="951ace80-f656-4e64-85e7-d4d54d793159" containerName="nova-manage" Dec 01 15:55:46 crc kubenswrapper[4739]: E1201 15:55:46.500049 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdac5301-3c64-40a8-85b3-67a92b22fa2a" containerName="nova-api-api" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.500060 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdac5301-3c64-40a8-85b3-67a92b22fa2a" containerName="nova-api-api" Dec 01 15:55:46 crc kubenswrapper[4739]: E1201 15:55:46.500084 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdac5301-3c64-40a8-85b3-67a92b22fa2a" containerName="nova-api-log" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.500095 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdac5301-3c64-40a8-85b3-67a92b22fa2a" containerName="nova-api-log" Dec 01 15:55:46 crc kubenswrapper[4739]: E1201 15:55:46.500119 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb53ff5e-3cee-4686-befd-d071117f4541" containerName="init" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.500129 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb53ff5e-3cee-4686-befd-d071117f4541" containerName="init" Dec 01 15:55:46 crc kubenswrapper[4739]: E1201 15:55:46.500152 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb53ff5e-3cee-4686-befd-d071117f4541" containerName="dnsmasq-dns" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.500162 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb53ff5e-3cee-4686-befd-d071117f4541" containerName="dnsmasq-dns" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.500481 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb53ff5e-3cee-4686-befd-d071117f4541" containerName="dnsmasq-dns" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.500506 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdac5301-3c64-40a8-85b3-67a92b22fa2a" containerName="nova-api-log" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.500517 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdac5301-3c64-40a8-85b3-67a92b22fa2a" containerName="nova-api-api" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.500530 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="951ace80-f656-4e64-85e7-d4d54d793159" containerName="nova-manage" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.501945 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.504062 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.504137 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.504213 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.511014 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.641651 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vh6p\" (UniqueName: \"kubernetes.io/projected/31d2c216-973d-47da-ae9b-2cdaa7f375ee-kube-api-access-4vh6p\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.641704 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31d2c216-973d-47da-ae9b-2cdaa7f375ee-config-data\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.641724 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31d2c216-973d-47da-ae9b-2cdaa7f375ee-public-tls-certs\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.641753 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31d2c216-973d-47da-ae9b-2cdaa7f375ee-internal-tls-certs\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.641802 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31d2c216-973d-47da-ae9b-2cdaa7f375ee-logs\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.641837 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31d2c216-973d-47da-ae9b-2cdaa7f375ee-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.743193 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vh6p\" (UniqueName: \"kubernetes.io/projected/31d2c216-973d-47da-ae9b-2cdaa7f375ee-kube-api-access-4vh6p\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.743254 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31d2c216-973d-47da-ae9b-2cdaa7f375ee-config-data\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.743273 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31d2c216-973d-47da-ae9b-2cdaa7f375ee-public-tls-certs\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.743304 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31d2c216-973d-47da-ae9b-2cdaa7f375ee-internal-tls-certs\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.743351 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31d2c216-973d-47da-ae9b-2cdaa7f375ee-logs\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.743384 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31d2c216-973d-47da-ae9b-2cdaa7f375ee-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.745957 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31d2c216-973d-47da-ae9b-2cdaa7f375ee-logs\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.747869 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31d2c216-973d-47da-ae9b-2cdaa7f375ee-internal-tls-certs\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.748287 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31d2c216-973d-47da-ae9b-2cdaa7f375ee-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.749034 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31d2c216-973d-47da-ae9b-2cdaa7f375ee-public-tls-certs\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.749137 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31d2c216-973d-47da-ae9b-2cdaa7f375ee-config-data\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.764368 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vh6p\" (UniqueName: \"kubernetes.io/projected/31d2c216-973d-47da-ae9b-2cdaa7f375ee-kube-api-access-4vh6p\") pod \"nova-api-0\" (UID: \"31d2c216-973d-47da-ae9b-2cdaa7f375ee\") " pod="openstack/nova-api-0" Dec 01 15:55:46 crc kubenswrapper[4739]: I1201 15:55:46.830404 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 15:55:47 crc kubenswrapper[4739]: I1201 15:55:47.289195 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 15:55:47 crc kubenswrapper[4739]: W1201 15:55:47.296628 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31d2c216_973d_47da_ae9b_2cdaa7f375ee.slice/crio-42e24bfe4dc52e42fe5500a74f8310fceaf529323a22afc5f58aff7f961b8add WatchSource:0}: Error finding container 42e24bfe4dc52e42fe5500a74f8310fceaf529323a22afc5f58aff7f961b8add: Status 404 returned error can't find the container with id 42e24bfe4dc52e42fe5500a74f8310fceaf529323a22afc5f58aff7f961b8add Dec 01 15:55:47 crc kubenswrapper[4739]: I1201 15:55:47.419089 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31d2c216-973d-47da-ae9b-2cdaa7f375ee","Type":"ContainerStarted","Data":"42e24bfe4dc52e42fe5500a74f8310fceaf529323a22afc5f58aff7f961b8add"} Dec 01 15:55:48 crc kubenswrapper[4739]: I1201 15:55:48.436237 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31d2c216-973d-47da-ae9b-2cdaa7f375ee","Type":"ContainerStarted","Data":"5b7291e265c3e48e66c6231f7dc9aef22b0fbea15ac833059bb26e894bdc09b7"} Dec 01 15:55:48 crc kubenswrapper[4739]: I1201 15:55:48.437680 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31d2c216-973d-47da-ae9b-2cdaa7f375ee","Type":"ContainerStarted","Data":"35545d27d53103e55f57cfd448197caefad0c0c3ce122864dc3c2793c313f7c7"} Dec 01 15:55:48 crc kubenswrapper[4739]: I1201 15:55:48.495744 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdac5301-3c64-40a8-85b3-67a92b22fa2a" path="/var/lib/kubelet/pods/fdac5301-3c64-40a8-85b3-67a92b22fa2a/volumes" Dec 01 15:55:48 crc kubenswrapper[4739]: I1201 15:55:48.709706 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b30c462a-2b55-4187-931e-12e78990741a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.175:8775/\": read tcp 10.217.0.2:52514->10.217.0.175:8775: read: connection reset by peer" Dec 01 15:55:48 crc kubenswrapper[4739]: I1201 15:55:48.710522 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b30c462a-2b55-4187-931e-12e78990741a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.175:8775/\": read tcp 10.217.0.2:52528->10.217.0.175:8775: read: connection reset by peer" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.375988 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.412704 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.412681037 podStartE2EDuration="3.412681037s" podCreationTimestamp="2025-12-01 15:55:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:55:48.459610363 +0000 UTC m=+1250.285356487" watchObservedRunningTime="2025-12-01 15:55:49.412681037 +0000 UTC m=+1251.238427131" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.451612 4739 generic.go:334] "Generic (PLEG): container finished" podID="b30c462a-2b55-4187-931e-12e78990741a" containerID="6f0e116dd42b56eca373c2ab2adb6c497716ffc059119fd760c3616cb644f343" exitCode=0 Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.452632 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.453586 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b30c462a-2b55-4187-931e-12e78990741a","Type":"ContainerDied","Data":"6f0e116dd42b56eca373c2ab2adb6c497716ffc059119fd760c3616cb644f343"} Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.453631 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b30c462a-2b55-4187-931e-12e78990741a","Type":"ContainerDied","Data":"e50d0eb90a7b608f5c60145475ec10f7fe624510c0a3eaf5dd92d0796548bd06"} Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.453652 4739 scope.go:117] "RemoveContainer" containerID="6f0e116dd42b56eca373c2ab2adb6c497716ffc059119fd760c3616cb644f343" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.488776 4739 scope.go:117] "RemoveContainer" containerID="7c67238b1af7b9e356f869c4c76a43339bb2f33a8d2452ecd5d77211e6a33247" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.498198 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-combined-ca-bundle\") pod \"b30c462a-2b55-4187-931e-12e78990741a\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.498295 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b30c462a-2b55-4187-931e-12e78990741a-logs\") pod \"b30c462a-2b55-4187-931e-12e78990741a\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.498402 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5dfq\" (UniqueName: \"kubernetes.io/projected/b30c462a-2b55-4187-931e-12e78990741a-kube-api-access-z5dfq\") pod \"b30c462a-2b55-4187-931e-12e78990741a\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.498508 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-nova-metadata-tls-certs\") pod \"b30c462a-2b55-4187-931e-12e78990741a\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.498682 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-config-data\") pod \"b30c462a-2b55-4187-931e-12e78990741a\" (UID: \"b30c462a-2b55-4187-931e-12e78990741a\") " Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.501369 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b30c462a-2b55-4187-931e-12e78990741a-logs" (OuterVolumeSpecName: "logs") pod "b30c462a-2b55-4187-931e-12e78990741a" (UID: "b30c462a-2b55-4187-931e-12e78990741a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.525654 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b30c462a-2b55-4187-931e-12e78990741a-kube-api-access-z5dfq" (OuterVolumeSpecName: "kube-api-access-z5dfq") pod "b30c462a-2b55-4187-931e-12e78990741a" (UID: "b30c462a-2b55-4187-931e-12e78990741a"). InnerVolumeSpecName "kube-api-access-z5dfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.541556 4739 scope.go:117] "RemoveContainer" containerID="6f0e116dd42b56eca373c2ab2adb6c497716ffc059119fd760c3616cb644f343" Dec 01 15:55:49 crc kubenswrapper[4739]: E1201 15:55:49.546243 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f0e116dd42b56eca373c2ab2adb6c497716ffc059119fd760c3616cb644f343\": container with ID starting with 6f0e116dd42b56eca373c2ab2adb6c497716ffc059119fd760c3616cb644f343 not found: ID does not exist" containerID="6f0e116dd42b56eca373c2ab2adb6c497716ffc059119fd760c3616cb644f343" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.546298 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f0e116dd42b56eca373c2ab2adb6c497716ffc059119fd760c3616cb644f343"} err="failed to get container status \"6f0e116dd42b56eca373c2ab2adb6c497716ffc059119fd760c3616cb644f343\": rpc error: code = NotFound desc = could not find container \"6f0e116dd42b56eca373c2ab2adb6c497716ffc059119fd760c3616cb644f343\": container with ID starting with 6f0e116dd42b56eca373c2ab2adb6c497716ffc059119fd760c3616cb644f343 not found: ID does not exist" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.546332 4739 scope.go:117] "RemoveContainer" containerID="7c67238b1af7b9e356f869c4c76a43339bb2f33a8d2452ecd5d77211e6a33247" Dec 01 15:55:49 crc kubenswrapper[4739]: E1201 15:55:49.546643 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c67238b1af7b9e356f869c4c76a43339bb2f33a8d2452ecd5d77211e6a33247\": container with ID starting with 7c67238b1af7b9e356f869c4c76a43339bb2f33a8d2452ecd5d77211e6a33247 not found: ID does not exist" containerID="7c67238b1af7b9e356f869c4c76a43339bb2f33a8d2452ecd5d77211e6a33247" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.546665 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c67238b1af7b9e356f869c4c76a43339bb2f33a8d2452ecd5d77211e6a33247"} err="failed to get container status \"7c67238b1af7b9e356f869c4c76a43339bb2f33a8d2452ecd5d77211e6a33247\": rpc error: code = NotFound desc = could not find container \"7c67238b1af7b9e356f869c4c76a43339bb2f33a8d2452ecd5d77211e6a33247\": container with ID starting with 7c67238b1af7b9e356f869c4c76a43339bb2f33a8d2452ecd5d77211e6a33247 not found: ID does not exist" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.549271 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-config-data" (OuterVolumeSpecName: "config-data") pod "b30c462a-2b55-4187-931e-12e78990741a" (UID: "b30c462a-2b55-4187-931e-12e78990741a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.558695 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b30c462a-2b55-4187-931e-12e78990741a" (UID: "b30c462a-2b55-4187-931e-12e78990741a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.574119 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "b30c462a-2b55-4187-931e-12e78990741a" (UID: "b30c462a-2b55-4187-931e-12e78990741a"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.601222 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.601246 4739 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b30c462a-2b55-4187-931e-12e78990741a-logs\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.601362 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5dfq\" (UniqueName: \"kubernetes.io/projected/b30c462a-2b55-4187-931e-12e78990741a-kube-api-access-z5dfq\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.601528 4739 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.601558 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b30c462a-2b55-4187-931e-12e78990741a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.795628 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.802805 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.815648 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:55:49 crc kubenswrapper[4739]: E1201 15:55:49.816002 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b30c462a-2b55-4187-931e-12e78990741a" containerName="nova-metadata-log" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.816019 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="b30c462a-2b55-4187-931e-12e78990741a" containerName="nova-metadata-log" Dec 01 15:55:49 crc kubenswrapper[4739]: E1201 15:55:49.816041 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b30c462a-2b55-4187-931e-12e78990741a" containerName="nova-metadata-metadata" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.816047 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="b30c462a-2b55-4187-931e-12e78990741a" containerName="nova-metadata-metadata" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.816200 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="b30c462a-2b55-4187-931e-12e78990741a" containerName="nova-metadata-metadata" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.816219 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="b30c462a-2b55-4187-931e-12e78990741a" containerName="nova-metadata-log" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.817579 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.820070 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.823709 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.835036 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.905691 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9-logs\") pod \"nova-metadata-0\" (UID: \"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9\") " pod="openstack/nova-metadata-0" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.905769 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9-config-data\") pod \"nova-metadata-0\" (UID: \"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9\") " pod="openstack/nova-metadata-0" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.905897 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kczz\" (UniqueName: \"kubernetes.io/projected/0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9-kube-api-access-7kczz\") pod \"nova-metadata-0\" (UID: \"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9\") " pod="openstack/nova-metadata-0" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.906018 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9\") " pod="openstack/nova-metadata-0" Dec 01 15:55:49 crc kubenswrapper[4739]: I1201 15:55:49.906084 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9\") " pod="openstack/nova-metadata-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.007492 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9-logs\") pod \"nova-metadata-0\" (UID: \"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9\") " pod="openstack/nova-metadata-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.007557 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9-config-data\") pod \"nova-metadata-0\" (UID: \"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9\") " pod="openstack/nova-metadata-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.007583 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kczz\" (UniqueName: \"kubernetes.io/projected/0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9-kube-api-access-7kczz\") pod \"nova-metadata-0\" (UID: \"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9\") " pod="openstack/nova-metadata-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.007621 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9\") " pod="openstack/nova-metadata-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.007651 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9\") " pod="openstack/nova-metadata-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.010864 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9\") " pod="openstack/nova-metadata-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.011100 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9-logs\") pod \"nova-metadata-0\" (UID: \"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9\") " pod="openstack/nova-metadata-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.014774 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9-config-data\") pod \"nova-metadata-0\" (UID: \"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9\") " pod="openstack/nova-metadata-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.028135 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9\") " pod="openstack/nova-metadata-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.062136 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kczz\" (UniqueName: \"kubernetes.io/projected/0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9-kube-api-access-7kczz\") pod \"nova-metadata-0\" (UID: \"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9\") " pod="openstack/nova-metadata-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.240664 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.268164 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.316891 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a928221-a9c0-4602-81d2-f0dede25dafa-combined-ca-bundle\") pod \"7a928221-a9c0-4602-81d2-f0dede25dafa\" (UID: \"7a928221-a9c0-4602-81d2-f0dede25dafa\") " Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.317006 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tm8m7\" (UniqueName: \"kubernetes.io/projected/7a928221-a9c0-4602-81d2-f0dede25dafa-kube-api-access-tm8m7\") pod \"7a928221-a9c0-4602-81d2-f0dede25dafa\" (UID: \"7a928221-a9c0-4602-81d2-f0dede25dafa\") " Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.317077 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a928221-a9c0-4602-81d2-f0dede25dafa-config-data\") pod \"7a928221-a9c0-4602-81d2-f0dede25dafa\" (UID: \"7a928221-a9c0-4602-81d2-f0dede25dafa\") " Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.322702 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a928221-a9c0-4602-81d2-f0dede25dafa-kube-api-access-tm8m7" (OuterVolumeSpecName: "kube-api-access-tm8m7") pod "7a928221-a9c0-4602-81d2-f0dede25dafa" (UID: "7a928221-a9c0-4602-81d2-f0dede25dafa"). InnerVolumeSpecName "kube-api-access-tm8m7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.346784 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a928221-a9c0-4602-81d2-f0dede25dafa-config-data" (OuterVolumeSpecName: "config-data") pod "7a928221-a9c0-4602-81d2-f0dede25dafa" (UID: "7a928221-a9c0-4602-81d2-f0dede25dafa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.380191 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a928221-a9c0-4602-81d2-f0dede25dafa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a928221-a9c0-4602-81d2-f0dede25dafa" (UID: "7a928221-a9c0-4602-81d2-f0dede25dafa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.419022 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tm8m7\" (UniqueName: \"kubernetes.io/projected/7a928221-a9c0-4602-81d2-f0dede25dafa-kube-api-access-tm8m7\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.419049 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a928221-a9c0-4602-81d2-f0dede25dafa-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.419059 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a928221-a9c0-4602-81d2-f0dede25dafa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.465830 4739 generic.go:334] "Generic (PLEG): container finished" podID="7a928221-a9c0-4602-81d2-f0dede25dafa" containerID="ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d" exitCode=0 Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.465891 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.465920 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7a928221-a9c0-4602-81d2-f0dede25dafa","Type":"ContainerDied","Data":"ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d"} Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.466856 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7a928221-a9c0-4602-81d2-f0dede25dafa","Type":"ContainerDied","Data":"ab5dc501fa1784d99cee69fe61ef0e746973eea7aa90ccc86a4f3ae438b7cefa"} Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.466878 4739 scope.go:117] "RemoveContainer" containerID="ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.487202 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b30c462a-2b55-4187-931e-12e78990741a" path="/var/lib/kubelet/pods/b30c462a-2b55-4187-931e-12e78990741a/volumes" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.493037 4739 scope.go:117] "RemoveContainer" containerID="ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d" Dec 01 15:55:50 crc kubenswrapper[4739]: E1201 15:55:50.493459 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d\": container with ID starting with ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d not found: ID does not exist" containerID="ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.493504 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d"} err="failed to get container status \"ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d\": rpc error: code = NotFound desc = could not find container \"ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d\": container with ID starting with ba0de1acf0f992210bfccb23b350fa51e22a89c72f92748cee2cb8d08b03314d not found: ID does not exist" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.512386 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.531657 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.548579 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 15:55:50 crc kubenswrapper[4739]: E1201 15:55:50.549323 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a928221-a9c0-4602-81d2-f0dede25dafa" containerName="nova-scheduler-scheduler" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.549349 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a928221-a9c0-4602-81d2-f0dede25dafa" containerName="nova-scheduler-scheduler" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.549763 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a928221-a9c0-4602-81d2-f0dede25dafa" containerName="nova-scheduler-scheduler" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.550860 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.554746 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.561782 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.625866 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bscss\" (UniqueName: \"kubernetes.io/projected/25515423-047d-445c-bae8-f54adf3ce705-kube-api-access-bscss\") pod \"nova-scheduler-0\" (UID: \"25515423-047d-445c-bae8-f54adf3ce705\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.626042 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25515423-047d-445c-bae8-f54adf3ce705-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"25515423-047d-445c-bae8-f54adf3ce705\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.626233 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25515423-047d-445c-bae8-f54adf3ce705-config-data\") pod \"nova-scheduler-0\" (UID: \"25515423-047d-445c-bae8-f54adf3ce705\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.687409 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 15:55:50 crc kubenswrapper[4739]: W1201 15:55:50.690141 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e21a1fb_5536_4eb8_8dd5_eeea64d1f6f9.slice/crio-4a0ec2e4fb9ae9eca989878568ce4459c6e9a2fa8e792c0776ba304f27c47ae7 WatchSource:0}: Error finding container 4a0ec2e4fb9ae9eca989878568ce4459c6e9a2fa8e792c0776ba304f27c47ae7: Status 404 returned error can't find the container with id 4a0ec2e4fb9ae9eca989878568ce4459c6e9a2fa8e792c0776ba304f27c47ae7 Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.727626 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25515423-047d-445c-bae8-f54adf3ce705-config-data\") pod \"nova-scheduler-0\" (UID: \"25515423-047d-445c-bae8-f54adf3ce705\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.727759 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bscss\" (UniqueName: \"kubernetes.io/projected/25515423-047d-445c-bae8-f54adf3ce705-kube-api-access-bscss\") pod \"nova-scheduler-0\" (UID: \"25515423-047d-445c-bae8-f54adf3ce705\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.727828 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25515423-047d-445c-bae8-f54adf3ce705-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"25515423-047d-445c-bae8-f54adf3ce705\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.734823 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25515423-047d-445c-bae8-f54adf3ce705-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"25515423-047d-445c-bae8-f54adf3ce705\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.735975 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25515423-047d-445c-bae8-f54adf3ce705-config-data\") pod \"nova-scheduler-0\" (UID: \"25515423-047d-445c-bae8-f54adf3ce705\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.745625 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bscss\" (UniqueName: \"kubernetes.io/projected/25515423-047d-445c-bae8-f54adf3ce705-kube-api-access-bscss\") pod \"nova-scheduler-0\" (UID: \"25515423-047d-445c-bae8-f54adf3ce705\") " pod="openstack/nova-scheduler-0" Dec 01 15:55:50 crc kubenswrapper[4739]: I1201 15:55:50.864554 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 15:55:51 crc kubenswrapper[4739]: W1201 15:55:51.380345 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25515423_047d_445c_bae8_f54adf3ce705.slice/crio-8ba652b019696e5632b90e43c38c17313cb7da0b018276e45bbeb14e7170df29 WatchSource:0}: Error finding container 8ba652b019696e5632b90e43c38c17313cb7da0b018276e45bbeb14e7170df29: Status 404 returned error can't find the container with id 8ba652b019696e5632b90e43c38c17313cb7da0b018276e45bbeb14e7170df29 Dec 01 15:55:51 crc kubenswrapper[4739]: I1201 15:55:51.403568 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 15:55:51 crc kubenswrapper[4739]: I1201 15:55:51.476522 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"25515423-047d-445c-bae8-f54adf3ce705","Type":"ContainerStarted","Data":"8ba652b019696e5632b90e43c38c17313cb7da0b018276e45bbeb14e7170df29"} Dec 01 15:55:51 crc kubenswrapper[4739]: I1201 15:55:51.480268 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9","Type":"ContainerStarted","Data":"aa11dd6a5f42124cfc3ca88778a3e55a10380e7960f9a578f766d625c0782664"} Dec 01 15:55:51 crc kubenswrapper[4739]: I1201 15:55:51.480321 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9","Type":"ContainerStarted","Data":"ff0e4c7f0dc41c19b10383673777e19ec52a3c6c6335ed998edd8dec301877f3"} Dec 01 15:55:51 crc kubenswrapper[4739]: I1201 15:55:51.480332 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9","Type":"ContainerStarted","Data":"4a0ec2e4fb9ae9eca989878568ce4459c6e9a2fa8e792c0776ba304f27c47ae7"} Dec 01 15:55:52 crc kubenswrapper[4739]: I1201 15:55:52.486111 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a928221-a9c0-4602-81d2-f0dede25dafa" path="/var/lib/kubelet/pods/7a928221-a9c0-4602-81d2-f0dede25dafa/volumes" Dec 01 15:55:52 crc kubenswrapper[4739]: I1201 15:55:52.490601 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"25515423-047d-445c-bae8-f54adf3ce705","Type":"ContainerStarted","Data":"7274903ac5d51f32fb850bd02bb65a4fcbe70d605c1e0e909bef4bc8415699d0"} Dec 01 15:55:52 crc kubenswrapper[4739]: I1201 15:55:52.507668 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.5076400899999998 podStartE2EDuration="3.50764009s" podCreationTimestamp="2025-12-01 15:55:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:55:51.497844149 +0000 UTC m=+1253.323590243" watchObservedRunningTime="2025-12-01 15:55:52.50764009 +0000 UTC m=+1254.333386214" Dec 01 15:55:52 crc kubenswrapper[4739]: I1201 15:55:52.512161 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.512146429 podStartE2EDuration="2.512146429s" podCreationTimestamp="2025-12-01 15:55:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:55:52.504329347 +0000 UTC m=+1254.330075521" watchObservedRunningTime="2025-12-01 15:55:52.512146429 +0000 UTC m=+1254.337892563" Dec 01 15:55:55 crc kubenswrapper[4739]: I1201 15:55:55.241070 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 15:55:55 crc kubenswrapper[4739]: I1201 15:55:55.241399 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 15:55:55 crc kubenswrapper[4739]: I1201 15:55:55.865138 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 15:55:56 crc kubenswrapper[4739]: I1201 15:55:56.832299 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 15:55:56 crc kubenswrapper[4739]: I1201 15:55:56.833508 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 15:55:57 crc kubenswrapper[4739]: I1201 15:55:57.847694 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="31d2c216-973d-47da-ae9b-2cdaa7f375ee" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.185:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 15:55:57 crc kubenswrapper[4739]: I1201 15:55:57.847709 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="31d2c216-973d-47da-ae9b-2cdaa7f375ee" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 15:56:00 crc kubenswrapper[4739]: I1201 15:56:00.241577 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 15:56:00 crc kubenswrapper[4739]: I1201 15:56:00.241914 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 15:56:00 crc kubenswrapper[4739]: I1201 15:56:00.865814 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 15:56:00 crc kubenswrapper[4739]: I1201 15:56:00.894774 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 15:56:01 crc kubenswrapper[4739]: I1201 15:56:01.258785 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.186:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 15:56:01 crc kubenswrapper[4739]: I1201 15:56:01.258811 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.186:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 15:56:01 crc kubenswrapper[4739]: I1201 15:56:01.619103 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 15:56:03 crc kubenswrapper[4739]: I1201 15:56:03.639519 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 15:56:06 crc kubenswrapper[4739]: I1201 15:56:06.839970 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 15:56:06 crc kubenswrapper[4739]: I1201 15:56:06.840750 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 15:56:06 crc kubenswrapper[4739]: I1201 15:56:06.842471 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 15:56:06 crc kubenswrapper[4739]: I1201 15:56:06.842518 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 15:56:06 crc kubenswrapper[4739]: I1201 15:56:06.851600 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 15:56:06 crc kubenswrapper[4739]: I1201 15:56:06.852912 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 15:56:09 crc kubenswrapper[4739]: I1201 15:56:09.621871 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:56:09 crc kubenswrapper[4739]: I1201 15:56:09.622277 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:56:10 crc kubenswrapper[4739]: I1201 15:56:10.255636 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 15:56:10 crc kubenswrapper[4739]: I1201 15:56:10.255712 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 15:56:10 crc kubenswrapper[4739]: I1201 15:56:10.264060 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 15:56:10 crc kubenswrapper[4739]: I1201 15:56:10.267900 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 15:56:18 crc kubenswrapper[4739]: I1201 15:56:18.158701 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 15:56:19 crc kubenswrapper[4739]: I1201 15:56:19.135493 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 15:56:22 crc kubenswrapper[4739]: I1201 15:56:22.772780 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="0f128863-acb7-4080-829a-01bd78766374" containerName="rabbitmq" containerID="cri-o://40c19226330b8e279fe364bce8b752d5af236a3696e9b11962cef197f1425812" gracePeriod=604796 Dec 01 15:56:23 crc kubenswrapper[4739]: I1201 15:56:23.483818 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="a2606100-66af-4acd-98a5-2edcfb281544" containerName="rabbitmq" containerID="cri-o://44d5f06f08742ef999ca64710212d1783cbaa21433efd216736620fff2acdb03" gracePeriod=604796 Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.560100 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.724636 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvk9f\" (UniqueName: \"kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-kube-api-access-fvk9f\") pod \"0f128863-acb7-4080-829a-01bd78766374\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.724685 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-server-conf\") pod \"0f128863-acb7-4080-829a-01bd78766374\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.724753 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-tls\") pod \"0f128863-acb7-4080-829a-01bd78766374\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.724771 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"0f128863-acb7-4080-829a-01bd78766374\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.724818 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f128863-acb7-4080-829a-01bd78766374-pod-info\") pod \"0f128863-acb7-4080-829a-01bd78766374\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.724878 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-erlang-cookie\") pod \"0f128863-acb7-4080-829a-01bd78766374\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.724924 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f128863-acb7-4080-829a-01bd78766374-erlang-cookie-secret\") pod \"0f128863-acb7-4080-829a-01bd78766374\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.724952 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-confd\") pod \"0f128863-acb7-4080-829a-01bd78766374\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.724999 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-plugins-conf\") pod \"0f128863-acb7-4080-829a-01bd78766374\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.725024 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-config-data\") pod \"0f128863-acb7-4080-829a-01bd78766374\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.725043 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-plugins\") pod \"0f128863-acb7-4080-829a-01bd78766374\" (UID: \"0f128863-acb7-4080-829a-01bd78766374\") " Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.725932 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0f128863-acb7-4080-829a-01bd78766374" (UID: "0f128863-acb7-4080-829a-01bd78766374"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.728149 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0f128863-acb7-4080-829a-01bd78766374" (UID: "0f128863-acb7-4080-829a-01bd78766374"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.728259 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0f128863-acb7-4080-829a-01bd78766374" (UID: "0f128863-acb7-4080-829a-01bd78766374"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.731866 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-kube-api-access-fvk9f" (OuterVolumeSpecName: "kube-api-access-fvk9f") pod "0f128863-acb7-4080-829a-01bd78766374" (UID: "0f128863-acb7-4080-829a-01bd78766374"). InnerVolumeSpecName "kube-api-access-fvk9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.734236 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f128863-acb7-4080-829a-01bd78766374-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0f128863-acb7-4080-829a-01bd78766374" (UID: "0f128863-acb7-4080-829a-01bd78766374"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.734335 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0f128863-acb7-4080-829a-01bd78766374-pod-info" (OuterVolumeSpecName: "pod-info") pod "0f128863-acb7-4080-829a-01bd78766374" (UID: "0f128863-acb7-4080-829a-01bd78766374"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.739814 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0f128863-acb7-4080-829a-01bd78766374" (UID: "0f128863-acb7-4080-829a-01bd78766374"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.750474 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "0f128863-acb7-4080-829a-01bd78766374" (UID: "0f128863-acb7-4080-829a-01bd78766374"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.761587 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-config-data" (OuterVolumeSpecName: "config-data") pod "0f128863-acb7-4080-829a-01bd78766374" (UID: "0f128863-acb7-4080-829a-01bd78766374"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.811696 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-server-conf" (OuterVolumeSpecName: "server-conf") pod "0f128863-acb7-4080-829a-01bd78766374" (UID: "0f128863-acb7-4080-829a-01bd78766374"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.826789 4739 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.826868 4739 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.826881 4739 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f128863-acb7-4080-829a-01bd78766374-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.826893 4739 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.828125 4739 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f128863-acb7-4080-829a-01bd78766374-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.828151 4739 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.828163 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.828174 4739 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.828186 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvk9f\" (UniqueName: \"kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-kube-api-access-fvk9f\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.828197 4739 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f128863-acb7-4080-829a-01bd78766374-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.850714 4739 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.852520 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0f128863-acb7-4080-829a-01bd78766374" (UID: "0f128863-acb7-4080-829a-01bd78766374"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.901623 4739 generic.go:334] "Generic (PLEG): container finished" podID="0f128863-acb7-4080-829a-01bd78766374" containerID="40c19226330b8e279fe364bce8b752d5af236a3696e9b11962cef197f1425812" exitCode=0 Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.901697 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.901715 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f128863-acb7-4080-829a-01bd78766374","Type":"ContainerDied","Data":"40c19226330b8e279fe364bce8b752d5af236a3696e9b11962cef197f1425812"} Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.902192 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f128863-acb7-4080-829a-01bd78766374","Type":"ContainerDied","Data":"38694cde5eeb8be42ab1b55e61dfbe17b18565af43593c8d0a588ef65ee37e97"} Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.902220 4739 scope.go:117] "RemoveContainer" containerID="40c19226330b8e279fe364bce8b752d5af236a3696e9b11962cef197f1425812" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.908504 4739 generic.go:334] "Generic (PLEG): container finished" podID="a2606100-66af-4acd-98a5-2edcfb281544" containerID="44d5f06f08742ef999ca64710212d1783cbaa21433efd216736620fff2acdb03" exitCode=0 Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.908579 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a2606100-66af-4acd-98a5-2edcfb281544","Type":"ContainerDied","Data":"44d5f06f08742ef999ca64710212d1783cbaa21433efd216736620fff2acdb03"} Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.929549 4739 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.929586 4739 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f128863-acb7-4080-829a-01bd78766374-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.936068 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.942364 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.943875 4739 scope.go:117] "RemoveContainer" containerID="17bbc7387e706b89896707fd45f1fbe1e75abe86f809cbdf894a0b9ec75c4520" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.972649 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 15:56:29 crc kubenswrapper[4739]: E1201 15:56:29.973323 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f128863-acb7-4080-829a-01bd78766374" containerName="rabbitmq" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.973335 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f128863-acb7-4080-829a-01bd78766374" containerName="rabbitmq" Dec 01 15:56:29 crc kubenswrapper[4739]: E1201 15:56:29.973349 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f128863-acb7-4080-829a-01bd78766374" containerName="setup-container" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.973356 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f128863-acb7-4080-829a-01bd78766374" containerName="setup-container" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.973570 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f128863-acb7-4080-829a-01bd78766374" containerName="rabbitmq" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.974499 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.978042 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.978175 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.978200 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-h8p42" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.978595 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.982508 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.987956 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.987960 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 01 15:56:29 crc kubenswrapper[4739]: I1201 15:56:29.993960 4739 scope.go:117] "RemoveContainer" containerID="40c19226330b8e279fe364bce8b752d5af236a3696e9b11962cef197f1425812" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.000283 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 15:56:30 crc kubenswrapper[4739]: E1201 15:56:30.000816 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40c19226330b8e279fe364bce8b752d5af236a3696e9b11962cef197f1425812\": container with ID starting with 40c19226330b8e279fe364bce8b752d5af236a3696e9b11962cef197f1425812 not found: ID does not exist" containerID="40c19226330b8e279fe364bce8b752d5af236a3696e9b11962cef197f1425812" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.000847 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40c19226330b8e279fe364bce8b752d5af236a3696e9b11962cef197f1425812"} err="failed to get container status \"40c19226330b8e279fe364bce8b752d5af236a3696e9b11962cef197f1425812\": rpc error: code = NotFound desc = could not find container \"40c19226330b8e279fe364bce8b752d5af236a3696e9b11962cef197f1425812\": container with ID starting with 40c19226330b8e279fe364bce8b752d5af236a3696e9b11962cef197f1425812 not found: ID does not exist" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.000871 4739 scope.go:117] "RemoveContainer" containerID="17bbc7387e706b89896707fd45f1fbe1e75abe86f809cbdf894a0b9ec75c4520" Dec 01 15:56:30 crc kubenswrapper[4739]: E1201 15:56:30.001983 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17bbc7387e706b89896707fd45f1fbe1e75abe86f809cbdf894a0b9ec75c4520\": container with ID starting with 17bbc7387e706b89896707fd45f1fbe1e75abe86f809cbdf894a0b9ec75c4520 not found: ID does not exist" containerID="17bbc7387e706b89896707fd45f1fbe1e75abe86f809cbdf894a0b9ec75c4520" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.002004 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17bbc7387e706b89896707fd45f1fbe1e75abe86f809cbdf894a0b9ec75c4520"} err="failed to get container status \"17bbc7387e706b89896707fd45f1fbe1e75abe86f809cbdf894a0b9ec75c4520\": rpc error: code = NotFound desc = could not find container \"17bbc7387e706b89896707fd45f1fbe1e75abe86f809cbdf894a0b9ec75c4520\": container with ID starting with 17bbc7387e706b89896707fd45f1fbe1e75abe86f809cbdf894a0b9ec75c4520 not found: ID does not exist" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.099525 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.133322 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1ad80536-1ee5-4fcd-9de2-16b7683f1870-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.133380 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1ad80536-1ee5-4fcd-9de2-16b7683f1870-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.133472 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1ad80536-1ee5-4fcd-9de2-16b7683f1870-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.133494 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.133674 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrz4f\" (UniqueName: \"kubernetes.io/projected/1ad80536-1ee5-4fcd-9de2-16b7683f1870-kube-api-access-xrz4f\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.133766 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1ad80536-1ee5-4fcd-9de2-16b7683f1870-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.133813 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1ad80536-1ee5-4fcd-9de2-16b7683f1870-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.133938 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1ad80536-1ee5-4fcd-9de2-16b7683f1870-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.133986 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1ad80536-1ee5-4fcd-9de2-16b7683f1870-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.134075 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1ad80536-1ee5-4fcd-9de2-16b7683f1870-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.134118 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ad80536-1ee5-4fcd-9de2-16b7683f1870-config-data\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.235846 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-server-conf\") pod \"a2606100-66af-4acd-98a5-2edcfb281544\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.235920 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-config-data\") pod \"a2606100-66af-4acd-98a5-2edcfb281544\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236007 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2606100-66af-4acd-98a5-2edcfb281544-pod-info\") pod \"a2606100-66af-4acd-98a5-2edcfb281544\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236073 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-erlang-cookie\") pod \"a2606100-66af-4acd-98a5-2edcfb281544\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236155 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2606100-66af-4acd-98a5-2edcfb281544-erlang-cookie-secret\") pod \"a2606100-66af-4acd-98a5-2edcfb281544\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236178 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-plugins\") pod \"a2606100-66af-4acd-98a5-2edcfb281544\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236207 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6fq8\" (UniqueName: \"kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-kube-api-access-s6fq8\") pod \"a2606100-66af-4acd-98a5-2edcfb281544\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236237 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-plugins-conf\") pod \"a2606100-66af-4acd-98a5-2edcfb281544\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236264 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-confd\") pod \"a2606100-66af-4acd-98a5-2edcfb281544\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236300 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-tls\") pod \"a2606100-66af-4acd-98a5-2edcfb281544\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236329 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"a2606100-66af-4acd-98a5-2edcfb281544\" (UID: \"a2606100-66af-4acd-98a5-2edcfb281544\") " Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236595 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1ad80536-1ee5-4fcd-9de2-16b7683f1870-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236640 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236674 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrz4f\" (UniqueName: \"kubernetes.io/projected/1ad80536-1ee5-4fcd-9de2-16b7683f1870-kube-api-access-xrz4f\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236714 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1ad80536-1ee5-4fcd-9de2-16b7683f1870-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236743 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1ad80536-1ee5-4fcd-9de2-16b7683f1870-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236749 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a2606100-66af-4acd-98a5-2edcfb281544" (UID: "a2606100-66af-4acd-98a5-2edcfb281544"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236816 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1ad80536-1ee5-4fcd-9de2-16b7683f1870-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236839 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1ad80536-1ee5-4fcd-9de2-16b7683f1870-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236870 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1ad80536-1ee5-4fcd-9de2-16b7683f1870-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236891 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ad80536-1ee5-4fcd-9de2-16b7683f1870-config-data\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236924 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1ad80536-1ee5-4fcd-9de2-16b7683f1870-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.236964 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1ad80536-1ee5-4fcd-9de2-16b7683f1870-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.237024 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a2606100-66af-4acd-98a5-2edcfb281544" (UID: "a2606100-66af-4acd-98a5-2edcfb281544"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.237109 4739 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.237431 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1ad80536-1ee5-4fcd-9de2-16b7683f1870-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.239553 4739 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.239584 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a2606100-66af-4acd-98a5-2edcfb281544" (UID: "a2606100-66af-4acd-98a5-2edcfb281544"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.241643 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ad80536-1ee5-4fcd-9de2-16b7683f1870-config-data\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.241688 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1ad80536-1ee5-4fcd-9de2-16b7683f1870-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.241846 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1ad80536-1ee5-4fcd-9de2-16b7683f1870-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.242766 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2606100-66af-4acd-98a5-2edcfb281544-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a2606100-66af-4acd-98a5-2edcfb281544" (UID: "a2606100-66af-4acd-98a5-2edcfb281544"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.243321 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1ad80536-1ee5-4fcd-9de2-16b7683f1870-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.243827 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a2606100-66af-4acd-98a5-2edcfb281544-pod-info" (OuterVolumeSpecName: "pod-info") pod "a2606100-66af-4acd-98a5-2edcfb281544" (UID: "a2606100-66af-4acd-98a5-2edcfb281544"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.244071 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1ad80536-1ee5-4fcd-9de2-16b7683f1870-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.245309 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-kube-api-access-s6fq8" (OuterVolumeSpecName: "kube-api-access-s6fq8") pod "a2606100-66af-4acd-98a5-2edcfb281544" (UID: "a2606100-66af-4acd-98a5-2edcfb281544"). InnerVolumeSpecName "kube-api-access-s6fq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.245840 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "a2606100-66af-4acd-98a5-2edcfb281544" (UID: "a2606100-66af-4acd-98a5-2edcfb281544"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.245939 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1ad80536-1ee5-4fcd-9de2-16b7683f1870-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.250989 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1ad80536-1ee5-4fcd-9de2-16b7683f1870-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.263248 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1ad80536-1ee5-4fcd-9de2-16b7683f1870-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.266697 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a2606100-66af-4acd-98a5-2edcfb281544" (UID: "a2606100-66af-4acd-98a5-2edcfb281544"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.277235 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrz4f\" (UniqueName: \"kubernetes.io/projected/1ad80536-1ee5-4fcd-9de2-16b7683f1870-kube-api-access-xrz4f\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.302274 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-server-conf" (OuterVolumeSpecName: "server-conf") pod "a2606100-66af-4acd-98a5-2edcfb281544" (UID: "a2606100-66af-4acd-98a5-2edcfb281544"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.310062 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-config-data" (OuterVolumeSpecName: "config-data") pod "a2606100-66af-4acd-98a5-2edcfb281544" (UID: "a2606100-66af-4acd-98a5-2edcfb281544"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.315252 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"1ad80536-1ee5-4fcd-9de2-16b7683f1870\") " pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.339294 4739 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.339322 4739 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2606100-66af-4acd-98a5-2edcfb281544-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.339347 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6fq8\" (UniqueName: \"kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-kube-api-access-s6fq8\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.339356 4739 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.339365 4739 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.339388 4739 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.339396 4739 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.339426 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2606100-66af-4acd-98a5-2edcfb281544-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.339434 4739 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2606100-66af-4acd-98a5-2edcfb281544-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.367222 4739 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.397544 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a2606100-66af-4acd-98a5-2edcfb281544" (UID: "a2606100-66af-4acd-98a5-2edcfb281544"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.440807 4739 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.440835 4739 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2606100-66af-4acd-98a5-2edcfb281544-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.487237 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f128863-acb7-4080-829a-01bd78766374" path="/var/lib/kubelet/pods/0f128863-acb7-4080-829a-01bd78766374/volumes" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.597666 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.925643 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-clhtt"] Dec 01 15:56:30 crc kubenswrapper[4739]: E1201 15:56:30.926373 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2606100-66af-4acd-98a5-2edcfb281544" containerName="setup-container" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.926389 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2606100-66af-4acd-98a5-2edcfb281544" containerName="setup-container" Dec 01 15:56:30 crc kubenswrapper[4739]: E1201 15:56:30.926408 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2606100-66af-4acd-98a5-2edcfb281544" containerName="rabbitmq" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.926435 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2606100-66af-4acd-98a5-2edcfb281544" containerName="rabbitmq" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.926845 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2606100-66af-4acd-98a5-2edcfb281544" containerName="rabbitmq" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.927784 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.929552 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.934448 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a2606100-66af-4acd-98a5-2edcfb281544","Type":"ContainerDied","Data":"8fcc7a9df5ee3f0cac5d68f0dcac1353400ab87cf57d8f6a7d69317a9d253f9a"} Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.934547 4739 scope.go:117] "RemoveContainer" containerID="44d5f06f08742ef999ca64710212d1783cbaa21433efd216736620fff2acdb03" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.934564 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.955517 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-clhtt"] Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.961423 4739 scope.go:117] "RemoveContainer" containerID="184f65dcda1a366c8645ddf60dcc32681e541fb0943f2bcdab6db7840d07555c" Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.982767 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 15:56:30 crc kubenswrapper[4739]: I1201 15:56:30.991297 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.008713 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.010283 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.013273 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.015253 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.015394 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.015625 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.015753 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.015874 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-x2bjw" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.015982 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.017870 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.071257 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9zjz\" (UniqueName: \"kubernetes.io/projected/53c5cc70-ab90-488b-ac95-4c20a4e840e2-kube-api-access-v9zjz\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.071305 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.071402 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-config\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.071480 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.071508 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.071544 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.138966 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 15:56:31 crc kubenswrapper[4739]: W1201 15:56:31.142511 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ad80536_1ee5_4fcd_9de2_16b7683f1870.slice/crio-5f8b2a3afb3280dec6c790f46d89d2c9c8818e8a1979c54aec672733854eee0e WatchSource:0}: Error finding container 5f8b2a3afb3280dec6c790f46d89d2c9c8818e8a1979c54aec672733854eee0e: Status 404 returned error can't find the container with id 5f8b2a3afb3280dec6c790f46d89d2c9c8818e8a1979c54aec672733854eee0e Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.175989 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176282 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176304 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9zjz\" (UniqueName: \"kubernetes.io/projected/53c5cc70-ab90-488b-ac95-4c20a4e840e2-kube-api-access-v9zjz\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176369 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176395 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-config\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176411 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176449 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176468 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176487 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176504 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxqjj\" (UniqueName: \"kubernetes.io/projected/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-kube-api-access-kxqjj\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176531 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176548 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176566 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176582 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176602 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176625 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.176642 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.177155 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-ovsdbserver-nb\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.177299 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-ovsdbserver-sb\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.177854 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-openstack-edpm-ipam\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.178076 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-config\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.179179 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-dns-svc\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.203509 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9zjz\" (UniqueName: \"kubernetes.io/projected/53c5cc70-ab90-488b-ac95-4c20a4e840e2-kube-api-access-v9zjz\") pod \"dnsmasq-dns-6447ccbd8f-clhtt\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.248031 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.278644 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.278846 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.278877 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.278907 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.278929 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.278952 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.278969 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxqjj\" (UniqueName: \"kubernetes.io/projected/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-kube-api-access-kxqjj\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.278997 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.279016 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.279035 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.279055 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.280304 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.280586 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.280923 4739 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.282616 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.282725 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.282790 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.284012 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.285738 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.308034 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.313197 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.318548 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxqjj\" (UniqueName: \"kubernetes.io/projected/c4dbaaa4-ac29-4909-9200-6db94e30f4e5-kube-api-access-kxqjj\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.328428 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4dbaaa4-ac29-4909-9200-6db94e30f4e5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.395885 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.717490 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 15:56:31 crc kubenswrapper[4739]: W1201 15:56:31.721733 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4dbaaa4_ac29_4909_9200_6db94e30f4e5.slice/crio-7eba71fd44bc3832519131b48211295393cb1149129bd60868f872307dc62459 WatchSource:0}: Error finding container 7eba71fd44bc3832519131b48211295393cb1149129bd60868f872307dc62459: Status 404 returned error can't find the container with id 7eba71fd44bc3832519131b48211295393cb1149129bd60868f872307dc62459 Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.845715 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-clhtt"] Dec 01 15:56:31 crc kubenswrapper[4739]: W1201 15:56:31.848118 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53c5cc70_ab90_488b_ac95_4c20a4e840e2.slice/crio-4bd7ac07df07afd54481fcf0b99dd23fdc5485c8da9c10f843f82e10f0c1816e WatchSource:0}: Error finding container 4bd7ac07df07afd54481fcf0b99dd23fdc5485c8da9c10f843f82e10f0c1816e: Status 404 returned error can't find the container with id 4bd7ac07df07afd54481fcf0b99dd23fdc5485c8da9c10f843f82e10f0c1816e Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.946544 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1ad80536-1ee5-4fcd-9de2-16b7683f1870","Type":"ContainerStarted","Data":"5f8b2a3afb3280dec6c790f46d89d2c9c8818e8a1979c54aec672733854eee0e"} Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.947728 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" event={"ID":"53c5cc70-ab90-488b-ac95-4c20a4e840e2","Type":"ContainerStarted","Data":"4bd7ac07df07afd54481fcf0b99dd23fdc5485c8da9c10f843f82e10f0c1816e"} Dec 01 15:56:31 crc kubenswrapper[4739]: I1201 15:56:31.948918 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c4dbaaa4-ac29-4909-9200-6db94e30f4e5","Type":"ContainerStarted","Data":"7eba71fd44bc3832519131b48211295393cb1149129bd60868f872307dc62459"} Dec 01 15:56:32 crc kubenswrapper[4739]: I1201 15:56:32.497152 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2606100-66af-4acd-98a5-2edcfb281544" path="/var/lib/kubelet/pods/a2606100-66af-4acd-98a5-2edcfb281544/volumes" Dec 01 15:56:32 crc kubenswrapper[4739]: I1201 15:56:32.962495 4739 generic.go:334] "Generic (PLEG): container finished" podID="53c5cc70-ab90-488b-ac95-4c20a4e840e2" containerID="7c0a3fd6aea8477732d61486a4342d3a171b41b054718c78039ac7398f2a727c" exitCode=0 Dec 01 15:56:32 crc kubenswrapper[4739]: I1201 15:56:32.962637 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" event={"ID":"53c5cc70-ab90-488b-ac95-4c20a4e840e2","Type":"ContainerDied","Data":"7c0a3fd6aea8477732d61486a4342d3a171b41b054718c78039ac7398f2a727c"} Dec 01 15:56:32 crc kubenswrapper[4739]: I1201 15:56:32.967609 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1ad80536-1ee5-4fcd-9de2-16b7683f1870","Type":"ContainerStarted","Data":"0590b0e33d210860592b34d7135ce712ebde2104eeb2417fa9f59bec045939d7"} Dec 01 15:56:33 crc kubenswrapper[4739]: I1201 15:56:33.980352 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c4dbaaa4-ac29-4909-9200-6db94e30f4e5","Type":"ContainerStarted","Data":"971424c3f433aa0f49405e62d541b92c104f3431920d769ebb8ffad1d71a5c02"} Dec 01 15:56:33 crc kubenswrapper[4739]: I1201 15:56:33.985213 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" event={"ID":"53c5cc70-ab90-488b-ac95-4c20a4e840e2","Type":"ContainerStarted","Data":"7443f830ad1675a23f9152ad421ba62fc9f3810e3cf71b294f914aeb47676fd1"} Dec 01 15:56:33 crc kubenswrapper[4739]: I1201 15:56:33.985373 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:34 crc kubenswrapper[4739]: I1201 15:56:34.041855 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" podStartSLOduration=4.041828006 podStartE2EDuration="4.041828006s" podCreationTimestamp="2025-12-01 15:56:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:56:34.040105703 +0000 UTC m=+1295.865851807" watchObservedRunningTime="2025-12-01 15:56:34.041828006 +0000 UTC m=+1295.867574130" Dec 01 15:56:39 crc kubenswrapper[4739]: I1201 15:56:39.622152 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:56:39 crc kubenswrapper[4739]: I1201 15:56:39.622816 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:56:39 crc kubenswrapper[4739]: I1201 15:56:39.622895 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:56:39 crc kubenswrapper[4739]: I1201 15:56:39.624058 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"faad8f9bb664920c3ea11c12b9b5f870015e0ec81598417117215899a38ccfdb"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:56:39 crc kubenswrapper[4739]: I1201 15:56:39.624179 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://faad8f9bb664920c3ea11c12b9b5f870015e0ec81598417117215899a38ccfdb" gracePeriod=600 Dec 01 15:56:40 crc kubenswrapper[4739]: I1201 15:56:40.067542 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="faad8f9bb664920c3ea11c12b9b5f870015e0ec81598417117215899a38ccfdb" exitCode=0 Dec 01 15:56:40 crc kubenswrapper[4739]: I1201 15:56:40.067599 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"faad8f9bb664920c3ea11c12b9b5f870015e0ec81598417117215899a38ccfdb"} Dec 01 15:56:40 crc kubenswrapper[4739]: I1201 15:56:40.067843 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a"} Dec 01 15:56:40 crc kubenswrapper[4739]: I1201 15:56:40.067864 4739 scope.go:117] "RemoveContainer" containerID="d23f235659cb740130c9b9ee149b02d7de5733c24075029cc9884583e5415dad" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.249809 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.320345 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-9dldl"] Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.320678 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b856c5697-9dldl" podUID="501d57ed-62bc-452a-b0ae-12835a26cd67" containerName="dnsmasq-dns" containerID="cri-o://de16423d8cc93ba215ca89de38a21ace1ff64866de049b16ff82087653c2de36" gracePeriod=10 Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.537562 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-d6twm"] Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.539818 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.571594 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-d6twm"] Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.618836 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.618965 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.618996 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp8qp\" (UniqueName: \"kubernetes.io/projected/fdf75503-6233-48bc-9c4e-66bd2baa2d16-kube-api-access-qp8qp\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.619035 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-config\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.620155 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.620314 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.722562 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.722694 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.722798 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.722887 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.722943 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp8qp\" (UniqueName: \"kubernetes.io/projected/fdf75503-6233-48bc-9c4e-66bd2baa2d16-kube-api-access-qp8qp\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.722975 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-config\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.724154 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-config\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.724981 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-openstack-edpm-ipam\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.725228 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-ovsdbserver-nb\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.725769 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-ovsdbserver-sb\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.726041 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-dns-svc\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.752750 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp8qp\" (UniqueName: \"kubernetes.io/projected/fdf75503-6233-48bc-9c4e-66bd2baa2d16-kube-api-access-qp8qp\") pod \"dnsmasq-dns-864d5fc68c-d6twm\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.832299 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.886181 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.925520 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-ovsdbserver-sb\") pod \"501d57ed-62bc-452a-b0ae-12835a26cd67\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.925641 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-dns-svc\") pod \"501d57ed-62bc-452a-b0ae-12835a26cd67\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.925688 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smlh4\" (UniqueName: \"kubernetes.io/projected/501d57ed-62bc-452a-b0ae-12835a26cd67-kube-api-access-smlh4\") pod \"501d57ed-62bc-452a-b0ae-12835a26cd67\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.925725 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-ovsdbserver-nb\") pod \"501d57ed-62bc-452a-b0ae-12835a26cd67\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.925766 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-config\") pod \"501d57ed-62bc-452a-b0ae-12835a26cd67\" (UID: \"501d57ed-62bc-452a-b0ae-12835a26cd67\") " Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.936829 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/501d57ed-62bc-452a-b0ae-12835a26cd67-kube-api-access-smlh4" (OuterVolumeSpecName: "kube-api-access-smlh4") pod "501d57ed-62bc-452a-b0ae-12835a26cd67" (UID: "501d57ed-62bc-452a-b0ae-12835a26cd67"). InnerVolumeSpecName "kube-api-access-smlh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.989250 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "501d57ed-62bc-452a-b0ae-12835a26cd67" (UID: "501d57ed-62bc-452a-b0ae-12835a26cd67"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.990974 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-config" (OuterVolumeSpecName: "config") pod "501d57ed-62bc-452a-b0ae-12835a26cd67" (UID: "501d57ed-62bc-452a-b0ae-12835a26cd67"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:56:41 crc kubenswrapper[4739]: I1201 15:56:41.994206 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "501d57ed-62bc-452a-b0ae-12835a26cd67" (UID: "501d57ed-62bc-452a-b0ae-12835a26cd67"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.007698 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "501d57ed-62bc-452a-b0ae-12835a26cd67" (UID: "501d57ed-62bc-452a-b0ae-12835a26cd67"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.029494 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.029521 4739 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.029534 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smlh4\" (UniqueName: \"kubernetes.io/projected/501d57ed-62bc-452a-b0ae-12835a26cd67-kube-api-access-smlh4\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.029545 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.029555 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/501d57ed-62bc-452a-b0ae-12835a26cd67-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.102865 4739 generic.go:334] "Generic (PLEG): container finished" podID="501d57ed-62bc-452a-b0ae-12835a26cd67" containerID="de16423d8cc93ba215ca89de38a21ace1ff64866de049b16ff82087653c2de36" exitCode=0 Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.102931 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b856c5697-9dldl" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.102954 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-9dldl" event={"ID":"501d57ed-62bc-452a-b0ae-12835a26cd67","Type":"ContainerDied","Data":"de16423d8cc93ba215ca89de38a21ace1ff64866de049b16ff82087653c2de36"} Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.104554 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b856c5697-9dldl" event={"ID":"501d57ed-62bc-452a-b0ae-12835a26cd67","Type":"ContainerDied","Data":"23c8e234b2b3dbbe00cff6e4b7e02ebee4f5c2410cf93d1db208b36bd0ec0838"} Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.104588 4739 scope.go:117] "RemoveContainer" containerID="de16423d8cc93ba215ca89de38a21ace1ff64866de049b16ff82087653c2de36" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.137154 4739 scope.go:117] "RemoveContainer" containerID="c6e7168e2b20ac0f450cbbe8f5ea9083625ed28eafeb9b4edcf2439928d01057" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.138344 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-9dldl"] Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.146072 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b856c5697-9dldl"] Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.157913 4739 scope.go:117] "RemoveContainer" containerID="de16423d8cc93ba215ca89de38a21ace1ff64866de049b16ff82087653c2de36" Dec 01 15:56:42 crc kubenswrapper[4739]: E1201 15:56:42.161830 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de16423d8cc93ba215ca89de38a21ace1ff64866de049b16ff82087653c2de36\": container with ID starting with de16423d8cc93ba215ca89de38a21ace1ff64866de049b16ff82087653c2de36 not found: ID does not exist" containerID="de16423d8cc93ba215ca89de38a21ace1ff64866de049b16ff82087653c2de36" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.161888 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de16423d8cc93ba215ca89de38a21ace1ff64866de049b16ff82087653c2de36"} err="failed to get container status \"de16423d8cc93ba215ca89de38a21ace1ff64866de049b16ff82087653c2de36\": rpc error: code = NotFound desc = could not find container \"de16423d8cc93ba215ca89de38a21ace1ff64866de049b16ff82087653c2de36\": container with ID starting with de16423d8cc93ba215ca89de38a21ace1ff64866de049b16ff82087653c2de36 not found: ID does not exist" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.161922 4739 scope.go:117] "RemoveContainer" containerID="c6e7168e2b20ac0f450cbbe8f5ea9083625ed28eafeb9b4edcf2439928d01057" Dec 01 15:56:42 crc kubenswrapper[4739]: E1201 15:56:42.162273 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6e7168e2b20ac0f450cbbe8f5ea9083625ed28eafeb9b4edcf2439928d01057\": container with ID starting with c6e7168e2b20ac0f450cbbe8f5ea9083625ed28eafeb9b4edcf2439928d01057 not found: ID does not exist" containerID="c6e7168e2b20ac0f450cbbe8f5ea9083625ed28eafeb9b4edcf2439928d01057" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.162308 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6e7168e2b20ac0f450cbbe8f5ea9083625ed28eafeb9b4edcf2439928d01057"} err="failed to get container status \"c6e7168e2b20ac0f450cbbe8f5ea9083625ed28eafeb9b4edcf2439928d01057\": rpc error: code = NotFound desc = could not find container \"c6e7168e2b20ac0f450cbbe8f5ea9083625ed28eafeb9b4edcf2439928d01057\": container with ID starting with c6e7168e2b20ac0f450cbbe8f5ea9083625ed28eafeb9b4edcf2439928d01057 not found: ID does not exist" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.399928 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-d6twm"] Dec 01 15:56:42 crc kubenswrapper[4739]: W1201 15:56:42.401090 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdf75503_6233_48bc_9c4e_66bd2baa2d16.slice/crio-68ab754fb6d0cff8bc76d145125c542f428b6972d32c877291877c61a308af82 WatchSource:0}: Error finding container 68ab754fb6d0cff8bc76d145125c542f428b6972d32c877291877c61a308af82: Status 404 returned error can't find the container with id 68ab754fb6d0cff8bc76d145125c542f428b6972d32c877291877c61a308af82 Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.528388 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="501d57ed-62bc-452a-b0ae-12835a26cd67" path="/var/lib/kubelet/pods/501d57ed-62bc-452a-b0ae-12835a26cd67/volumes" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.686237 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985"] Dec 01 15:56:42 crc kubenswrapper[4739]: E1201 15:56:42.686641 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501d57ed-62bc-452a-b0ae-12835a26cd67" containerName="init" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.686670 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="501d57ed-62bc-452a-b0ae-12835a26cd67" containerName="init" Dec 01 15:56:42 crc kubenswrapper[4739]: E1201 15:56:42.686686 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501d57ed-62bc-452a-b0ae-12835a26cd67" containerName="dnsmasq-dns" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.686692 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="501d57ed-62bc-452a-b0ae-12835a26cd67" containerName="dnsmasq-dns" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.686852 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="501d57ed-62bc-452a-b0ae-12835a26cd67" containerName="dnsmasq-dns" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.687430 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.688996 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.689598 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.689771 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.692146 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.696853 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985"] Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.747658 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtmdj\" (UniqueName: \"kubernetes.io/projected/f7b6b83c-405e-4572-ba8f-85664c4a5aba-kube-api-access-vtmdj\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4p985\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.747713 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4p985\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.747844 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4p985\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.748013 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4p985\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.849278 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtmdj\" (UniqueName: \"kubernetes.io/projected/f7b6b83c-405e-4572-ba8f-85664c4a5aba-kube-api-access-vtmdj\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4p985\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.849321 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4p985\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.849378 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4p985\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.849472 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4p985\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.853976 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4p985\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.854124 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4p985\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.860556 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4p985\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:56:42 crc kubenswrapper[4739]: I1201 15:56:42.871024 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtmdj\" (UniqueName: \"kubernetes.io/projected/f7b6b83c-405e-4572-ba8f-85664c4a5aba-kube-api-access-vtmdj\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4p985\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:56:43 crc kubenswrapper[4739]: I1201 15:56:43.042656 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:56:43 crc kubenswrapper[4739]: I1201 15:56:43.138057 4739 generic.go:334] "Generic (PLEG): container finished" podID="fdf75503-6233-48bc-9c4e-66bd2baa2d16" containerID="fc24315c126dc12e60ab5ed5df5094ff8f67b44a60e97ec1864ecbaebec58eb5" exitCode=0 Dec 01 15:56:43 crc kubenswrapper[4739]: I1201 15:56:43.138869 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" event={"ID":"fdf75503-6233-48bc-9c4e-66bd2baa2d16","Type":"ContainerDied","Data":"fc24315c126dc12e60ab5ed5df5094ff8f67b44a60e97ec1864ecbaebec58eb5"} Dec 01 15:56:43 crc kubenswrapper[4739]: I1201 15:56:43.138917 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" event={"ID":"fdf75503-6233-48bc-9c4e-66bd2baa2d16","Type":"ContainerStarted","Data":"68ab754fb6d0cff8bc76d145125c542f428b6972d32c877291877c61a308af82"} Dec 01 15:56:43 crc kubenswrapper[4739]: I1201 15:56:43.693351 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985"] Dec 01 15:56:43 crc kubenswrapper[4739]: W1201 15:56:43.699162 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7b6b83c_405e_4572_ba8f_85664c4a5aba.slice/crio-9434e601e3ab6db28ef8da590b34358885263988000faa2a940d7cbaac30e0d5 WatchSource:0}: Error finding container 9434e601e3ab6db28ef8da590b34358885263988000faa2a940d7cbaac30e0d5: Status 404 returned error can't find the container with id 9434e601e3ab6db28ef8da590b34358885263988000faa2a940d7cbaac30e0d5 Dec 01 15:56:44 crc kubenswrapper[4739]: I1201 15:56:44.150853 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" event={"ID":"fdf75503-6233-48bc-9c4e-66bd2baa2d16","Type":"ContainerStarted","Data":"f1c382a69ee627f69cfddb0935265aec232c49ebd46b50f28b7152fac2d72c32"} Dec 01 15:56:44 crc kubenswrapper[4739]: I1201 15:56:44.151237 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:44 crc kubenswrapper[4739]: I1201 15:56:44.153023 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" event={"ID":"f7b6b83c-405e-4572-ba8f-85664c4a5aba","Type":"ContainerStarted","Data":"9434e601e3ab6db28ef8da590b34358885263988000faa2a940d7cbaac30e0d5"} Dec 01 15:56:44 crc kubenswrapper[4739]: I1201 15:56:44.177932 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" podStartSLOduration=3.177909367 podStartE2EDuration="3.177909367s" podCreationTimestamp="2025-12-01 15:56:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:56:44.174322226 +0000 UTC m=+1306.000068320" watchObservedRunningTime="2025-12-01 15:56:44.177909367 +0000 UTC m=+1306.003655461" Dec 01 15:56:51 crc kubenswrapper[4739]: I1201 15:56:51.887573 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 15:56:51 crc kubenswrapper[4739]: I1201 15:56:51.950132 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-clhtt"] Dec 01 15:56:51 crc kubenswrapper[4739]: I1201 15:56:51.950393 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" podUID="53c5cc70-ab90-488b-ac95-4c20a4e840e2" containerName="dnsmasq-dns" containerID="cri-o://7443f830ad1675a23f9152ad421ba62fc9f3810e3cf71b294f914aeb47676fd1" gracePeriod=10 Dec 01 15:56:52 crc kubenswrapper[4739]: I1201 15:56:52.250465 4739 generic.go:334] "Generic (PLEG): container finished" podID="53c5cc70-ab90-488b-ac95-4c20a4e840e2" containerID="7443f830ad1675a23f9152ad421ba62fc9f3810e3cf71b294f914aeb47676fd1" exitCode=0 Dec 01 15:56:52 crc kubenswrapper[4739]: I1201 15:56:52.250504 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" event={"ID":"53c5cc70-ab90-488b-ac95-4c20a4e840e2","Type":"ContainerDied","Data":"7443f830ad1675a23f9152ad421ba62fc9f3810e3cf71b294f914aeb47676fd1"} Dec 01 15:56:53 crc kubenswrapper[4739]: I1201 15:56:53.756740 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:53 crc kubenswrapper[4739]: I1201 15:56:53.919292 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-ovsdbserver-sb\") pod \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " Dec 01 15:56:53 crc kubenswrapper[4739]: I1201 15:56:53.919521 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-ovsdbserver-nb\") pod \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " Dec 01 15:56:53 crc kubenswrapper[4739]: I1201 15:56:53.919591 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-config\") pod \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " Dec 01 15:56:53 crc kubenswrapper[4739]: I1201 15:56:53.919654 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-openstack-edpm-ipam\") pod \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " Dec 01 15:56:53 crc kubenswrapper[4739]: I1201 15:56:53.919696 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9zjz\" (UniqueName: \"kubernetes.io/projected/53c5cc70-ab90-488b-ac95-4c20a4e840e2-kube-api-access-v9zjz\") pod \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " Dec 01 15:56:53 crc kubenswrapper[4739]: I1201 15:56:53.919786 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-dns-svc\") pod \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\" (UID: \"53c5cc70-ab90-488b-ac95-4c20a4e840e2\") " Dec 01 15:56:53 crc kubenswrapper[4739]: I1201 15:56:53.926539 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53c5cc70-ab90-488b-ac95-4c20a4e840e2-kube-api-access-v9zjz" (OuterVolumeSpecName: "kube-api-access-v9zjz") pod "53c5cc70-ab90-488b-ac95-4c20a4e840e2" (UID: "53c5cc70-ab90-488b-ac95-4c20a4e840e2"). InnerVolumeSpecName "kube-api-access-v9zjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:56:53 crc kubenswrapper[4739]: I1201 15:56:53.968279 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-config" (OuterVolumeSpecName: "config") pod "53c5cc70-ab90-488b-ac95-4c20a4e840e2" (UID: "53c5cc70-ab90-488b-ac95-4c20a4e840e2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:56:53 crc kubenswrapper[4739]: I1201 15:56:53.970877 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "53c5cc70-ab90-488b-ac95-4c20a4e840e2" (UID: "53c5cc70-ab90-488b-ac95-4c20a4e840e2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:56:53 crc kubenswrapper[4739]: I1201 15:56:53.977893 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "53c5cc70-ab90-488b-ac95-4c20a4e840e2" (UID: "53c5cc70-ab90-488b-ac95-4c20a4e840e2"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:56:53 crc kubenswrapper[4739]: I1201 15:56:53.984091 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "53c5cc70-ab90-488b-ac95-4c20a4e840e2" (UID: "53c5cc70-ab90-488b-ac95-4c20a4e840e2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:56:53 crc kubenswrapper[4739]: I1201 15:56:53.994857 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "53c5cc70-ab90-488b-ac95-4c20a4e840e2" (UID: "53c5cc70-ab90-488b-ac95-4c20a4e840e2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:56:54 crc kubenswrapper[4739]: I1201 15:56:54.022712 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9zjz\" (UniqueName: \"kubernetes.io/projected/53c5cc70-ab90-488b-ac95-4c20a4e840e2-kube-api-access-v9zjz\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:54 crc kubenswrapper[4739]: I1201 15:56:54.022762 4739 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:54 crc kubenswrapper[4739]: I1201 15:56:54.022773 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:54 crc kubenswrapper[4739]: I1201 15:56:54.022783 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:54 crc kubenswrapper[4739]: I1201 15:56:54.022792 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:54 crc kubenswrapper[4739]: I1201 15:56:54.022803 4739 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/53c5cc70-ab90-488b-ac95-4c20a4e840e2-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 15:56:54 crc kubenswrapper[4739]: I1201 15:56:54.268091 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" event={"ID":"f7b6b83c-405e-4572-ba8f-85664c4a5aba","Type":"ContainerStarted","Data":"bf5451c2fa5291fc98935f3660e6f7c55c1e9340b98cb227a0f4a27326da5e30"} Dec 01 15:56:54 crc kubenswrapper[4739]: I1201 15:56:54.270261 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" event={"ID":"53c5cc70-ab90-488b-ac95-4c20a4e840e2","Type":"ContainerDied","Data":"4bd7ac07df07afd54481fcf0b99dd23fdc5485c8da9c10f843f82e10f0c1816e"} Dec 01 15:56:54 crc kubenswrapper[4739]: I1201 15:56:54.270304 4739 scope.go:117] "RemoveContainer" containerID="7443f830ad1675a23f9152ad421ba62fc9f3810e3cf71b294f914aeb47676fd1" Dec 01 15:56:54 crc kubenswrapper[4739]: I1201 15:56:54.270335 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6447ccbd8f-clhtt" Dec 01 15:56:54 crc kubenswrapper[4739]: I1201 15:56:54.287457 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" podStartSLOduration=2.509106269 podStartE2EDuration="12.287439179s" podCreationTimestamp="2025-12-01 15:56:42 +0000 UTC" firstStartedPulling="2025-12-01 15:56:43.70103924 +0000 UTC m=+1305.526785334" lastFinishedPulling="2025-12-01 15:56:53.47937215 +0000 UTC m=+1315.305118244" observedRunningTime="2025-12-01 15:56:54.286687096 +0000 UTC m=+1316.112433190" watchObservedRunningTime="2025-12-01 15:56:54.287439179 +0000 UTC m=+1316.113185293" Dec 01 15:56:54 crc kubenswrapper[4739]: I1201 15:56:54.306246 4739 scope.go:117] "RemoveContainer" containerID="7c0a3fd6aea8477732d61486a4342d3a171b41b054718c78039ac7398f2a727c" Dec 01 15:56:54 crc kubenswrapper[4739]: I1201 15:56:54.313854 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-clhtt"] Dec 01 15:56:54 crc kubenswrapper[4739]: I1201 15:56:54.325859 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6447ccbd8f-clhtt"] Dec 01 15:56:54 crc kubenswrapper[4739]: I1201 15:56:54.488053 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53c5cc70-ab90-488b-ac95-4c20a4e840e2" path="/var/lib/kubelet/pods/53c5cc70-ab90-488b-ac95-4c20a4e840e2/volumes" Dec 01 15:57:05 crc kubenswrapper[4739]: I1201 15:57:05.379265 4739 generic.go:334] "Generic (PLEG): container finished" podID="1ad80536-1ee5-4fcd-9de2-16b7683f1870" containerID="0590b0e33d210860592b34d7135ce712ebde2104eeb2417fa9f59bec045939d7" exitCode=0 Dec 01 15:57:05 crc kubenswrapper[4739]: I1201 15:57:05.379374 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1ad80536-1ee5-4fcd-9de2-16b7683f1870","Type":"ContainerDied","Data":"0590b0e33d210860592b34d7135ce712ebde2104eeb2417fa9f59bec045939d7"} Dec 01 15:57:05 crc kubenswrapper[4739]: I1201 15:57:05.387868 4739 generic.go:334] "Generic (PLEG): container finished" podID="f7b6b83c-405e-4572-ba8f-85664c4a5aba" containerID="bf5451c2fa5291fc98935f3660e6f7c55c1e9340b98cb227a0f4a27326da5e30" exitCode=0 Dec 01 15:57:05 crc kubenswrapper[4739]: I1201 15:57:05.387933 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" event={"ID":"f7b6b83c-405e-4572-ba8f-85664c4a5aba","Type":"ContainerDied","Data":"bf5451c2fa5291fc98935f3660e6f7c55c1e9340b98cb227a0f4a27326da5e30"} Dec 01 15:57:06 crc kubenswrapper[4739]: I1201 15:57:06.407019 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1ad80536-1ee5-4fcd-9de2-16b7683f1870","Type":"ContainerStarted","Data":"719618a9d6a42dad579414925b57fa2f1f0f8b078ad38ef5b446440341a76ef4"} Dec 01 15:57:06 crc kubenswrapper[4739]: I1201 15:57:06.407635 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 01 15:57:06 crc kubenswrapper[4739]: I1201 15:57:06.443953 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.443934632 podStartE2EDuration="37.443934632s" podCreationTimestamp="2025-12-01 15:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:57:06.442955542 +0000 UTC m=+1328.268701636" watchObservedRunningTime="2025-12-01 15:57:06.443934632 +0000 UTC m=+1328.269680736" Dec 01 15:57:06 crc kubenswrapper[4739]: I1201 15:57:06.979401 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.089719 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-repo-setup-combined-ca-bundle\") pod \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.089805 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtmdj\" (UniqueName: \"kubernetes.io/projected/f7b6b83c-405e-4572-ba8f-85664c4a5aba-kube-api-access-vtmdj\") pod \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.089835 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-ssh-key\") pod \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.089969 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-inventory\") pod \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\" (UID: \"f7b6b83c-405e-4572-ba8f-85664c4a5aba\") " Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.096391 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "f7b6b83c-405e-4572-ba8f-85664c4a5aba" (UID: "f7b6b83c-405e-4572-ba8f-85664c4a5aba"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.101969 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7b6b83c-405e-4572-ba8f-85664c4a5aba-kube-api-access-vtmdj" (OuterVolumeSpecName: "kube-api-access-vtmdj") pod "f7b6b83c-405e-4572-ba8f-85664c4a5aba" (UID: "f7b6b83c-405e-4572-ba8f-85664c4a5aba"). InnerVolumeSpecName "kube-api-access-vtmdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.116091 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-inventory" (OuterVolumeSpecName: "inventory") pod "f7b6b83c-405e-4572-ba8f-85664c4a5aba" (UID: "f7b6b83c-405e-4572-ba8f-85664c4a5aba"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.141690 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f7b6b83c-405e-4572-ba8f-85664c4a5aba" (UID: "f7b6b83c-405e-4572-ba8f-85664c4a5aba"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.191993 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.192048 4739 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.192061 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtmdj\" (UniqueName: \"kubernetes.io/projected/f7b6b83c-405e-4572-ba8f-85664c4a5aba-kube-api-access-vtmdj\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.192074 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7b6b83c-405e-4572-ba8f-85664c4a5aba-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.419449 4739 generic.go:334] "Generic (PLEG): container finished" podID="c4dbaaa4-ac29-4909-9200-6db94e30f4e5" containerID="971424c3f433aa0f49405e62d541b92c104f3431920d769ebb8ffad1d71a5c02" exitCode=0 Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.419500 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c4dbaaa4-ac29-4909-9200-6db94e30f4e5","Type":"ContainerDied","Data":"971424c3f433aa0f49405e62d541b92c104f3431920d769ebb8ffad1d71a5c02"} Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.426536 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" event={"ID":"f7b6b83c-405e-4572-ba8f-85664c4a5aba","Type":"ContainerDied","Data":"9434e601e3ab6db28ef8da590b34358885263988000faa2a940d7cbaac30e0d5"} Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.426571 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.426584 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9434e601e3ab6db28ef8da590b34358885263988000faa2a940d7cbaac30e0d5" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.593657 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5"] Dec 01 15:57:07 crc kubenswrapper[4739]: E1201 15:57:07.594079 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c5cc70-ab90-488b-ac95-4c20a4e840e2" containerName="dnsmasq-dns" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.594093 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c5cc70-ab90-488b-ac95-4c20a4e840e2" containerName="dnsmasq-dns" Dec 01 15:57:07 crc kubenswrapper[4739]: E1201 15:57:07.594122 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c5cc70-ab90-488b-ac95-4c20a4e840e2" containerName="init" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.594131 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c5cc70-ab90-488b-ac95-4c20a4e840e2" containerName="init" Dec 01 15:57:07 crc kubenswrapper[4739]: E1201 15:57:07.594144 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7b6b83c-405e-4572-ba8f-85664c4a5aba" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.594155 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7b6b83c-405e-4572-ba8f-85664c4a5aba" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.594376 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="53c5cc70-ab90-488b-ac95-4c20a4e840e2" containerName="dnsmasq-dns" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.594396 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7b6b83c-405e-4572-ba8f-85664c4a5aba" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.596526 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.601278 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.601296 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.602965 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.603664 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.603878 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5"] Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.700226 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.700707 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.700767 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2krp2\" (UniqueName: \"kubernetes.io/projected/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-kube-api-access-2krp2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.700859 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.803071 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.803242 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.803284 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.803324 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2krp2\" (UniqueName: \"kubernetes.io/projected/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-kube-api-access-2krp2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.808808 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.810088 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.810267 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.832265 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2krp2\" (UniqueName: \"kubernetes.io/projected/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-kube-api-access-2krp2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 15:57:07 crc kubenswrapper[4739]: I1201 15:57:07.918924 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 15:57:08 crc kubenswrapper[4739]: I1201 15:57:08.439550 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c4dbaaa4-ac29-4909-9200-6db94e30f4e5","Type":"ContainerStarted","Data":"563cfca0d3469ca3f8d41e22275ae23519f968ba9e417c979413707a1661e9e7"} Dec 01 15:57:08 crc kubenswrapper[4739]: I1201 15:57:08.441063 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5"] Dec 01 15:57:08 crc kubenswrapper[4739]: I1201 15:57:08.441132 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:57:08 crc kubenswrapper[4739]: W1201 15:57:08.444510 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae86132a_1c2e_4e6b_9b3a_6f8b63be4270.slice/crio-5a602b5c47a3a35d80a9ec3eb576f31799f83e27015048af199421eea241b4e4 WatchSource:0}: Error finding container 5a602b5c47a3a35d80a9ec3eb576f31799f83e27015048af199421eea241b4e4: Status 404 returned error can't find the container with id 5a602b5c47a3a35d80a9ec3eb576f31799f83e27015048af199421eea241b4e4 Dec 01 15:57:08 crc kubenswrapper[4739]: I1201 15:57:08.513065 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.513047727 podStartE2EDuration="38.513047727s" podCreationTimestamp="2025-12-01 15:56:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:57:08.48430049 +0000 UTC m=+1330.310046664" watchObservedRunningTime="2025-12-01 15:57:08.513047727 +0000 UTC m=+1330.338793821" Dec 01 15:57:09 crc kubenswrapper[4739]: I1201 15:57:09.447618 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" event={"ID":"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270","Type":"ContainerStarted","Data":"cb26e7e043112190326b4e04beaaab20e5f5d7b71e3274844c6b90e1c92d063b"} Dec 01 15:57:09 crc kubenswrapper[4739]: I1201 15:57:09.447922 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" event={"ID":"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270","Type":"ContainerStarted","Data":"5a602b5c47a3a35d80a9ec3eb576f31799f83e27015048af199421eea241b4e4"} Dec 01 15:57:09 crc kubenswrapper[4739]: I1201 15:57:09.473604 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" podStartSLOduration=1.853314919 podStartE2EDuration="2.473587921s" podCreationTimestamp="2025-12-01 15:57:07 +0000 UTC" firstStartedPulling="2025-12-01 15:57:08.447306629 +0000 UTC m=+1330.273052723" lastFinishedPulling="2025-12-01 15:57:09.067579631 +0000 UTC m=+1330.893325725" observedRunningTime="2025-12-01 15:57:09.465580394 +0000 UTC m=+1331.291326478" watchObservedRunningTime="2025-12-01 15:57:09.473587921 +0000 UTC m=+1331.299334005" Dec 01 15:57:20 crc kubenswrapper[4739]: I1201 15:57:20.602203 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 01 15:57:21 crc kubenswrapper[4739]: I1201 15:57:21.400651 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:57:37 crc kubenswrapper[4739]: I1201 15:57:37.892433 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4j8nh"] Dec 01 15:57:37 crc kubenswrapper[4739]: I1201 15:57:37.900623 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:37 crc kubenswrapper[4739]: I1201 15:57:37.936178 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4j8nh"] Dec 01 15:57:37 crc kubenswrapper[4739]: I1201 15:57:37.958610 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/531c33d1-8219-41df-b673-c919e977ba21-catalog-content\") pod \"redhat-marketplace-4j8nh\" (UID: \"531c33d1-8219-41df-b673-c919e977ba21\") " pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:37 crc kubenswrapper[4739]: I1201 15:57:37.958817 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25mhj\" (UniqueName: \"kubernetes.io/projected/531c33d1-8219-41df-b673-c919e977ba21-kube-api-access-25mhj\") pod \"redhat-marketplace-4j8nh\" (UID: \"531c33d1-8219-41df-b673-c919e977ba21\") " pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:37 crc kubenswrapper[4739]: I1201 15:57:37.958850 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/531c33d1-8219-41df-b673-c919e977ba21-utilities\") pod \"redhat-marketplace-4j8nh\" (UID: \"531c33d1-8219-41df-b673-c919e977ba21\") " pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:38 crc kubenswrapper[4739]: I1201 15:57:38.059871 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25mhj\" (UniqueName: \"kubernetes.io/projected/531c33d1-8219-41df-b673-c919e977ba21-kube-api-access-25mhj\") pod \"redhat-marketplace-4j8nh\" (UID: \"531c33d1-8219-41df-b673-c919e977ba21\") " pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:38 crc kubenswrapper[4739]: I1201 15:57:38.060821 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/531c33d1-8219-41df-b673-c919e977ba21-utilities\") pod \"redhat-marketplace-4j8nh\" (UID: \"531c33d1-8219-41df-b673-c919e977ba21\") " pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:38 crc kubenswrapper[4739]: I1201 15:57:38.060927 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/531c33d1-8219-41df-b673-c919e977ba21-catalog-content\") pod \"redhat-marketplace-4j8nh\" (UID: \"531c33d1-8219-41df-b673-c919e977ba21\") " pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:38 crc kubenswrapper[4739]: I1201 15:57:38.061767 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/531c33d1-8219-41df-b673-c919e977ba21-utilities\") pod \"redhat-marketplace-4j8nh\" (UID: \"531c33d1-8219-41df-b673-c919e977ba21\") " pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:38 crc kubenswrapper[4739]: I1201 15:57:38.061943 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/531c33d1-8219-41df-b673-c919e977ba21-catalog-content\") pod \"redhat-marketplace-4j8nh\" (UID: \"531c33d1-8219-41df-b673-c919e977ba21\") " pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:38 crc kubenswrapper[4739]: I1201 15:57:38.082487 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25mhj\" (UniqueName: \"kubernetes.io/projected/531c33d1-8219-41df-b673-c919e977ba21-kube-api-access-25mhj\") pod \"redhat-marketplace-4j8nh\" (UID: \"531c33d1-8219-41df-b673-c919e977ba21\") " pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:38 crc kubenswrapper[4739]: I1201 15:57:38.250559 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:38 crc kubenswrapper[4739]: I1201 15:57:38.731630 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4j8nh"] Dec 01 15:57:38 crc kubenswrapper[4739]: W1201 15:57:38.740172 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod531c33d1_8219_41df_b673_c919e977ba21.slice/crio-c6561bb56e21e0b6cf55f426177ee6654be1e9b1337eaa4c1c1af6e53ecd4328 WatchSource:0}: Error finding container c6561bb56e21e0b6cf55f426177ee6654be1e9b1337eaa4c1c1af6e53ecd4328: Status 404 returned error can't find the container with id c6561bb56e21e0b6cf55f426177ee6654be1e9b1337eaa4c1c1af6e53ecd4328 Dec 01 15:57:38 crc kubenswrapper[4739]: I1201 15:57:38.782320 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4j8nh" event={"ID":"531c33d1-8219-41df-b673-c919e977ba21","Type":"ContainerStarted","Data":"c6561bb56e21e0b6cf55f426177ee6654be1e9b1337eaa4c1c1af6e53ecd4328"} Dec 01 15:57:39 crc kubenswrapper[4739]: I1201 15:57:39.793275 4739 generic.go:334] "Generic (PLEG): container finished" podID="531c33d1-8219-41df-b673-c919e977ba21" containerID="c1ea680b256e789094c493931e84fb1088b3ab8d1685e8d2b1b670569b68f43f" exitCode=0 Dec 01 15:57:39 crc kubenswrapper[4739]: I1201 15:57:39.793375 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4j8nh" event={"ID":"531c33d1-8219-41df-b673-c919e977ba21","Type":"ContainerDied","Data":"c1ea680b256e789094c493931e84fb1088b3ab8d1685e8d2b1b670569b68f43f"} Dec 01 15:57:41 crc kubenswrapper[4739]: I1201 15:57:41.822330 4739 generic.go:334] "Generic (PLEG): container finished" podID="531c33d1-8219-41df-b673-c919e977ba21" containerID="6b33de296fbdf571b784e8c0f6a2a83b1318cdbb05a9dc602327b1fa3fea448d" exitCode=0 Dec 01 15:57:41 crc kubenswrapper[4739]: I1201 15:57:41.822476 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4j8nh" event={"ID":"531c33d1-8219-41df-b673-c919e977ba21","Type":"ContainerDied","Data":"6b33de296fbdf571b784e8c0f6a2a83b1318cdbb05a9dc602327b1fa3fea448d"} Dec 01 15:57:42 crc kubenswrapper[4739]: I1201 15:57:42.841762 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4j8nh" event={"ID":"531c33d1-8219-41df-b673-c919e977ba21","Type":"ContainerStarted","Data":"ea5dca71800cb8cd4a24f6c846730bb3b77ee52a666b76fc9b9e6339aceb7a95"} Dec 01 15:57:42 crc kubenswrapper[4739]: I1201 15:57:42.864517 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4j8nh" podStartSLOduration=3.233674862 podStartE2EDuration="5.864500323s" podCreationTimestamp="2025-12-01 15:57:37 +0000 UTC" firstStartedPulling="2025-12-01 15:57:39.797579074 +0000 UTC m=+1361.623325208" lastFinishedPulling="2025-12-01 15:57:42.428404565 +0000 UTC m=+1364.254150669" observedRunningTime="2025-12-01 15:57:42.862742239 +0000 UTC m=+1364.688488403" watchObservedRunningTime="2025-12-01 15:57:42.864500323 +0000 UTC m=+1364.690246417" Dec 01 15:57:48 crc kubenswrapper[4739]: I1201 15:57:48.251332 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:48 crc kubenswrapper[4739]: I1201 15:57:48.252117 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:48 crc kubenswrapper[4739]: I1201 15:57:48.322887 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:48 crc kubenswrapper[4739]: I1201 15:57:48.993897 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:49 crc kubenswrapper[4739]: I1201 15:57:49.075581 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4j8nh"] Dec 01 15:57:50 crc kubenswrapper[4739]: I1201 15:57:50.930796 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4j8nh" podUID="531c33d1-8219-41df-b673-c919e977ba21" containerName="registry-server" containerID="cri-o://ea5dca71800cb8cd4a24f6c846730bb3b77ee52a666b76fc9b9e6339aceb7a95" gracePeriod=2 Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.466641 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.565693 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/531c33d1-8219-41df-b673-c919e977ba21-catalog-content\") pod \"531c33d1-8219-41df-b673-c919e977ba21\" (UID: \"531c33d1-8219-41df-b673-c919e977ba21\") " Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.565858 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/531c33d1-8219-41df-b673-c919e977ba21-utilities\") pod \"531c33d1-8219-41df-b673-c919e977ba21\" (UID: \"531c33d1-8219-41df-b673-c919e977ba21\") " Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.565882 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25mhj\" (UniqueName: \"kubernetes.io/projected/531c33d1-8219-41df-b673-c919e977ba21-kube-api-access-25mhj\") pod \"531c33d1-8219-41df-b673-c919e977ba21\" (UID: \"531c33d1-8219-41df-b673-c919e977ba21\") " Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.571867 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/531c33d1-8219-41df-b673-c919e977ba21-utilities" (OuterVolumeSpecName: "utilities") pod "531c33d1-8219-41df-b673-c919e977ba21" (UID: "531c33d1-8219-41df-b673-c919e977ba21"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.579694 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/531c33d1-8219-41df-b673-c919e977ba21-kube-api-access-25mhj" (OuterVolumeSpecName: "kube-api-access-25mhj") pod "531c33d1-8219-41df-b673-c919e977ba21" (UID: "531c33d1-8219-41df-b673-c919e977ba21"). InnerVolumeSpecName "kube-api-access-25mhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.593287 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/531c33d1-8219-41df-b673-c919e977ba21-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "531c33d1-8219-41df-b673-c919e977ba21" (UID: "531c33d1-8219-41df-b673-c919e977ba21"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.669370 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/531c33d1-8219-41df-b673-c919e977ba21-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.669472 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/531c33d1-8219-41df-b673-c919e977ba21-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.669488 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25mhj\" (UniqueName: \"kubernetes.io/projected/531c33d1-8219-41df-b673-c919e977ba21-kube-api-access-25mhj\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.948304 4739 generic.go:334] "Generic (PLEG): container finished" podID="531c33d1-8219-41df-b673-c919e977ba21" containerID="ea5dca71800cb8cd4a24f6c846730bb3b77ee52a666b76fc9b9e6339aceb7a95" exitCode=0 Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.948346 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4j8nh" event={"ID":"531c33d1-8219-41df-b673-c919e977ba21","Type":"ContainerDied","Data":"ea5dca71800cb8cd4a24f6c846730bb3b77ee52a666b76fc9b9e6339aceb7a95"} Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.948374 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4j8nh" event={"ID":"531c33d1-8219-41df-b673-c919e977ba21","Type":"ContainerDied","Data":"c6561bb56e21e0b6cf55f426177ee6654be1e9b1337eaa4c1c1af6e53ecd4328"} Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.948394 4739 scope.go:117] "RemoveContainer" containerID="ea5dca71800cb8cd4a24f6c846730bb3b77ee52a666b76fc9b9e6339aceb7a95" Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.948524 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4j8nh" Dec 01 15:57:51 crc kubenswrapper[4739]: I1201 15:57:51.990035 4739 scope.go:117] "RemoveContainer" containerID="6b33de296fbdf571b784e8c0f6a2a83b1318cdbb05a9dc602327b1fa3fea448d" Dec 01 15:57:52 crc kubenswrapper[4739]: I1201 15:57:52.033338 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4j8nh"] Dec 01 15:57:52 crc kubenswrapper[4739]: I1201 15:57:52.037079 4739 scope.go:117] "RemoveContainer" containerID="c1ea680b256e789094c493931e84fb1088b3ab8d1685e8d2b1b670569b68f43f" Dec 01 15:57:52 crc kubenswrapper[4739]: I1201 15:57:52.051304 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4j8nh"] Dec 01 15:57:52 crc kubenswrapper[4739]: I1201 15:57:52.099576 4739 scope.go:117] "RemoveContainer" containerID="ea5dca71800cb8cd4a24f6c846730bb3b77ee52a666b76fc9b9e6339aceb7a95" Dec 01 15:57:52 crc kubenswrapper[4739]: E1201 15:57:52.100597 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea5dca71800cb8cd4a24f6c846730bb3b77ee52a666b76fc9b9e6339aceb7a95\": container with ID starting with ea5dca71800cb8cd4a24f6c846730bb3b77ee52a666b76fc9b9e6339aceb7a95 not found: ID does not exist" containerID="ea5dca71800cb8cd4a24f6c846730bb3b77ee52a666b76fc9b9e6339aceb7a95" Dec 01 15:57:52 crc kubenswrapper[4739]: I1201 15:57:52.100677 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea5dca71800cb8cd4a24f6c846730bb3b77ee52a666b76fc9b9e6339aceb7a95"} err="failed to get container status \"ea5dca71800cb8cd4a24f6c846730bb3b77ee52a666b76fc9b9e6339aceb7a95\": rpc error: code = NotFound desc = could not find container \"ea5dca71800cb8cd4a24f6c846730bb3b77ee52a666b76fc9b9e6339aceb7a95\": container with ID starting with ea5dca71800cb8cd4a24f6c846730bb3b77ee52a666b76fc9b9e6339aceb7a95 not found: ID does not exist" Dec 01 15:57:52 crc kubenswrapper[4739]: I1201 15:57:52.100724 4739 scope.go:117] "RemoveContainer" containerID="6b33de296fbdf571b784e8c0f6a2a83b1318cdbb05a9dc602327b1fa3fea448d" Dec 01 15:57:52 crc kubenswrapper[4739]: E1201 15:57:52.101124 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b33de296fbdf571b784e8c0f6a2a83b1318cdbb05a9dc602327b1fa3fea448d\": container with ID starting with 6b33de296fbdf571b784e8c0f6a2a83b1318cdbb05a9dc602327b1fa3fea448d not found: ID does not exist" containerID="6b33de296fbdf571b784e8c0f6a2a83b1318cdbb05a9dc602327b1fa3fea448d" Dec 01 15:57:52 crc kubenswrapper[4739]: I1201 15:57:52.101158 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b33de296fbdf571b784e8c0f6a2a83b1318cdbb05a9dc602327b1fa3fea448d"} err="failed to get container status \"6b33de296fbdf571b784e8c0f6a2a83b1318cdbb05a9dc602327b1fa3fea448d\": rpc error: code = NotFound desc = could not find container \"6b33de296fbdf571b784e8c0f6a2a83b1318cdbb05a9dc602327b1fa3fea448d\": container with ID starting with 6b33de296fbdf571b784e8c0f6a2a83b1318cdbb05a9dc602327b1fa3fea448d not found: ID does not exist" Dec 01 15:57:52 crc kubenswrapper[4739]: I1201 15:57:52.101185 4739 scope.go:117] "RemoveContainer" containerID="c1ea680b256e789094c493931e84fb1088b3ab8d1685e8d2b1b670569b68f43f" Dec 01 15:57:52 crc kubenswrapper[4739]: E1201 15:57:52.101596 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1ea680b256e789094c493931e84fb1088b3ab8d1685e8d2b1b670569b68f43f\": container with ID starting with c1ea680b256e789094c493931e84fb1088b3ab8d1685e8d2b1b670569b68f43f not found: ID does not exist" containerID="c1ea680b256e789094c493931e84fb1088b3ab8d1685e8d2b1b670569b68f43f" Dec 01 15:57:52 crc kubenswrapper[4739]: I1201 15:57:52.101643 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1ea680b256e789094c493931e84fb1088b3ab8d1685e8d2b1b670569b68f43f"} err="failed to get container status \"c1ea680b256e789094c493931e84fb1088b3ab8d1685e8d2b1b670569b68f43f\": rpc error: code = NotFound desc = could not find container \"c1ea680b256e789094c493931e84fb1088b3ab8d1685e8d2b1b670569b68f43f\": container with ID starting with c1ea680b256e789094c493931e84fb1088b3ab8d1685e8d2b1b670569b68f43f not found: ID does not exist" Dec 01 15:57:52 crc kubenswrapper[4739]: I1201 15:57:52.503620 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="531c33d1-8219-41df-b673-c919e977ba21" path="/var/lib/kubelet/pods/531c33d1-8219-41df-b673-c919e977ba21/volumes" Dec 01 15:58:20 crc kubenswrapper[4739]: I1201 15:58:20.283312 4739 scope.go:117] "RemoveContainer" containerID="7c3c3f663d2c893ed61b1ce9e3eb2b38720420458e5eb4a3a399b87cb2d2c964" Dec 01 15:58:20 crc kubenswrapper[4739]: I1201 15:58:20.335548 4739 scope.go:117] "RemoveContainer" containerID="d0f94041a898118ba56ed8b833fda3241a2179e10a73ae2ca4a679e00f5d0128" Dec 01 15:58:25 crc kubenswrapper[4739]: I1201 15:58:25.938250 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z6dj6"] Dec 01 15:58:25 crc kubenswrapper[4739]: E1201 15:58:25.939885 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="531c33d1-8219-41df-b673-c919e977ba21" containerName="registry-server" Dec 01 15:58:25 crc kubenswrapper[4739]: I1201 15:58:25.939918 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="531c33d1-8219-41df-b673-c919e977ba21" containerName="registry-server" Dec 01 15:58:25 crc kubenswrapper[4739]: E1201 15:58:25.939952 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="531c33d1-8219-41df-b673-c919e977ba21" containerName="extract-content" Dec 01 15:58:25 crc kubenswrapper[4739]: I1201 15:58:25.939969 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="531c33d1-8219-41df-b673-c919e977ba21" containerName="extract-content" Dec 01 15:58:25 crc kubenswrapper[4739]: E1201 15:58:25.940006 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="531c33d1-8219-41df-b673-c919e977ba21" containerName="extract-utilities" Dec 01 15:58:25 crc kubenswrapper[4739]: I1201 15:58:25.940024 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="531c33d1-8219-41df-b673-c919e977ba21" containerName="extract-utilities" Dec 01 15:58:25 crc kubenswrapper[4739]: I1201 15:58:25.940574 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="531c33d1-8219-41df-b673-c919e977ba21" containerName="registry-server" Dec 01 15:58:25 crc kubenswrapper[4739]: I1201 15:58:25.943741 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6dj6" Dec 01 15:58:25 crc kubenswrapper[4739]: I1201 15:58:25.952482 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z6dj6"] Dec 01 15:58:26 crc kubenswrapper[4739]: I1201 15:58:26.049638 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/658c7094-c779-4d53-bd61-fc0b76e02934-catalog-content\") pod \"redhat-operators-z6dj6\" (UID: \"658c7094-c779-4d53-bd61-fc0b76e02934\") " pod="openshift-marketplace/redhat-operators-z6dj6" Dec 01 15:58:26 crc kubenswrapper[4739]: I1201 15:58:26.049934 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/658c7094-c779-4d53-bd61-fc0b76e02934-utilities\") pod \"redhat-operators-z6dj6\" (UID: \"658c7094-c779-4d53-bd61-fc0b76e02934\") " pod="openshift-marketplace/redhat-operators-z6dj6" Dec 01 15:58:26 crc kubenswrapper[4739]: I1201 15:58:26.050248 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsq4j\" (UniqueName: \"kubernetes.io/projected/658c7094-c779-4d53-bd61-fc0b76e02934-kube-api-access-hsq4j\") pod \"redhat-operators-z6dj6\" (UID: \"658c7094-c779-4d53-bd61-fc0b76e02934\") " pod="openshift-marketplace/redhat-operators-z6dj6" Dec 01 15:58:26 crc kubenswrapper[4739]: I1201 15:58:26.152758 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsq4j\" (UniqueName: \"kubernetes.io/projected/658c7094-c779-4d53-bd61-fc0b76e02934-kube-api-access-hsq4j\") pod \"redhat-operators-z6dj6\" (UID: \"658c7094-c779-4d53-bd61-fc0b76e02934\") " pod="openshift-marketplace/redhat-operators-z6dj6" Dec 01 15:58:26 crc kubenswrapper[4739]: I1201 15:58:26.152932 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/658c7094-c779-4d53-bd61-fc0b76e02934-catalog-content\") pod \"redhat-operators-z6dj6\" (UID: \"658c7094-c779-4d53-bd61-fc0b76e02934\") " pod="openshift-marketplace/redhat-operators-z6dj6" Dec 01 15:58:26 crc kubenswrapper[4739]: I1201 15:58:26.152977 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/658c7094-c779-4d53-bd61-fc0b76e02934-utilities\") pod \"redhat-operators-z6dj6\" (UID: \"658c7094-c779-4d53-bd61-fc0b76e02934\") " pod="openshift-marketplace/redhat-operators-z6dj6" Dec 01 15:58:26 crc kubenswrapper[4739]: I1201 15:58:26.153523 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/658c7094-c779-4d53-bd61-fc0b76e02934-catalog-content\") pod \"redhat-operators-z6dj6\" (UID: \"658c7094-c779-4d53-bd61-fc0b76e02934\") " pod="openshift-marketplace/redhat-operators-z6dj6" Dec 01 15:58:26 crc kubenswrapper[4739]: I1201 15:58:26.153737 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/658c7094-c779-4d53-bd61-fc0b76e02934-utilities\") pod \"redhat-operators-z6dj6\" (UID: \"658c7094-c779-4d53-bd61-fc0b76e02934\") " pod="openshift-marketplace/redhat-operators-z6dj6" Dec 01 15:58:26 crc kubenswrapper[4739]: I1201 15:58:26.185746 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsq4j\" (UniqueName: \"kubernetes.io/projected/658c7094-c779-4d53-bd61-fc0b76e02934-kube-api-access-hsq4j\") pod \"redhat-operators-z6dj6\" (UID: \"658c7094-c779-4d53-bd61-fc0b76e02934\") " pod="openshift-marketplace/redhat-operators-z6dj6" Dec 01 15:58:26 crc kubenswrapper[4739]: I1201 15:58:26.277208 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6dj6" Dec 01 15:58:26 crc kubenswrapper[4739]: I1201 15:58:26.707010 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z6dj6"] Dec 01 15:58:27 crc kubenswrapper[4739]: I1201 15:58:27.357923 4739 generic.go:334] "Generic (PLEG): container finished" podID="658c7094-c779-4d53-bd61-fc0b76e02934" containerID="c7eee3d1d48634e84fdabad133e0c95f837d0a76fc30c7182f1e8ab599070a7c" exitCode=0 Dec 01 15:58:27 crc kubenswrapper[4739]: I1201 15:58:27.358010 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6dj6" event={"ID":"658c7094-c779-4d53-bd61-fc0b76e02934","Type":"ContainerDied","Data":"c7eee3d1d48634e84fdabad133e0c95f837d0a76fc30c7182f1e8ab599070a7c"} Dec 01 15:58:27 crc kubenswrapper[4739]: I1201 15:58:27.358172 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6dj6" event={"ID":"658c7094-c779-4d53-bd61-fc0b76e02934","Type":"ContainerStarted","Data":"989f5426d2174a56eeb759bd77e5ce7f90d53029b392603ebcdee5906e7e2487"} Dec 01 15:58:36 crc kubenswrapper[4739]: I1201 15:58:36.465850 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6dj6" event={"ID":"658c7094-c779-4d53-bd61-fc0b76e02934","Type":"ContainerStarted","Data":"c3b3661123a5aeeae7bd6ab456ae57c0791d2cd6f1a48efb8314cf951ad874fd"} Dec 01 15:58:39 crc kubenswrapper[4739]: I1201 15:58:39.518927 4739 generic.go:334] "Generic (PLEG): container finished" podID="658c7094-c779-4d53-bd61-fc0b76e02934" containerID="c3b3661123a5aeeae7bd6ab456ae57c0791d2cd6f1a48efb8314cf951ad874fd" exitCode=0 Dec 01 15:58:39 crc kubenswrapper[4739]: I1201 15:58:39.518996 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6dj6" event={"ID":"658c7094-c779-4d53-bd61-fc0b76e02934","Type":"ContainerDied","Data":"c3b3661123a5aeeae7bd6ab456ae57c0791d2cd6f1a48efb8314cf951ad874fd"} Dec 01 15:58:39 crc kubenswrapper[4739]: I1201 15:58:39.622115 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:58:39 crc kubenswrapper[4739]: I1201 15:58:39.622217 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:58:41 crc kubenswrapper[4739]: I1201 15:58:41.550238 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6dj6" event={"ID":"658c7094-c779-4d53-bd61-fc0b76e02934","Type":"ContainerStarted","Data":"c02de5308eff5ab2226028b8a5d56603c75b7a00aece9e2db31483891e19ca87"} Dec 01 15:58:41 crc kubenswrapper[4739]: I1201 15:58:41.601662 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z6dj6" podStartSLOduration=3.593189973 podStartE2EDuration="16.601630249s" podCreationTimestamp="2025-12-01 15:58:25 +0000 UTC" firstStartedPulling="2025-12-01 15:58:27.361407359 +0000 UTC m=+1409.187153443" lastFinishedPulling="2025-12-01 15:58:40.369847625 +0000 UTC m=+1422.195593719" observedRunningTime="2025-12-01 15:58:41.581181577 +0000 UTC m=+1423.406927711" watchObservedRunningTime="2025-12-01 15:58:41.601630249 +0000 UTC m=+1423.427376353" Dec 01 15:58:46 crc kubenswrapper[4739]: I1201 15:58:46.278408 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z6dj6" Dec 01 15:58:46 crc kubenswrapper[4739]: I1201 15:58:46.278773 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z6dj6" Dec 01 15:58:47 crc kubenswrapper[4739]: I1201 15:58:47.324542 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z6dj6" podUID="658c7094-c779-4d53-bd61-fc0b76e02934" containerName="registry-server" probeResult="failure" output=< Dec 01 15:58:47 crc kubenswrapper[4739]: timeout: failed to connect service ":50051" within 1s Dec 01 15:58:47 crc kubenswrapper[4739]: > Dec 01 15:58:56 crc kubenswrapper[4739]: I1201 15:58:56.340518 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z6dj6" Dec 01 15:58:56 crc kubenswrapper[4739]: I1201 15:58:56.398682 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z6dj6" Dec 01 15:58:56 crc kubenswrapper[4739]: I1201 15:58:56.956640 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z6dj6"] Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.122642 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fnwbc"] Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.123295 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fnwbc" podUID="434fbbdb-10c8-461e-812f-7725b27b0111" containerName="registry-server" containerID="cri-o://dedcf54eafbcff7957421b1569934a5e4a9c7137a06b70c8c026ebfd4e63a614" gracePeriod=2 Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.554004 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.624903 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgqw5\" (UniqueName: \"kubernetes.io/projected/434fbbdb-10c8-461e-812f-7725b27b0111-kube-api-access-kgqw5\") pod \"434fbbdb-10c8-461e-812f-7725b27b0111\" (UID: \"434fbbdb-10c8-461e-812f-7725b27b0111\") " Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.625028 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434fbbdb-10c8-461e-812f-7725b27b0111-catalog-content\") pod \"434fbbdb-10c8-461e-812f-7725b27b0111\" (UID: \"434fbbdb-10c8-461e-812f-7725b27b0111\") " Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.625060 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434fbbdb-10c8-461e-812f-7725b27b0111-utilities\") pod \"434fbbdb-10c8-461e-812f-7725b27b0111\" (UID: \"434fbbdb-10c8-461e-812f-7725b27b0111\") " Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.627785 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/434fbbdb-10c8-461e-812f-7725b27b0111-utilities" (OuterVolumeSpecName: "utilities") pod "434fbbdb-10c8-461e-812f-7725b27b0111" (UID: "434fbbdb-10c8-461e-812f-7725b27b0111"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.648706 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/434fbbdb-10c8-461e-812f-7725b27b0111-kube-api-access-kgqw5" (OuterVolumeSpecName: "kube-api-access-kgqw5") pod "434fbbdb-10c8-461e-812f-7725b27b0111" (UID: "434fbbdb-10c8-461e-812f-7725b27b0111"). InnerVolumeSpecName "kube-api-access-kgqw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.718246 4739 generic.go:334] "Generic (PLEG): container finished" podID="434fbbdb-10c8-461e-812f-7725b27b0111" containerID="dedcf54eafbcff7957421b1569934a5e4a9c7137a06b70c8c026ebfd4e63a614" exitCode=0 Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.718391 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fnwbc" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.719276 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fnwbc" event={"ID":"434fbbdb-10c8-461e-812f-7725b27b0111","Type":"ContainerDied","Data":"dedcf54eafbcff7957421b1569934a5e4a9c7137a06b70c8c026ebfd4e63a614"} Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.719315 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fnwbc" event={"ID":"434fbbdb-10c8-461e-812f-7725b27b0111","Type":"ContainerDied","Data":"a1ca7b89dddc1a98f6adc160d57374f4628af2f72b0cf7a3853f6b2e7dd4d217"} Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.719338 4739 scope.go:117] "RemoveContainer" containerID="dedcf54eafbcff7957421b1569934a5e4a9c7137a06b70c8c026ebfd4e63a614" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.727796 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434fbbdb-10c8-461e-812f-7725b27b0111-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.727839 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgqw5\" (UniqueName: \"kubernetes.io/projected/434fbbdb-10c8-461e-812f-7725b27b0111-kube-api-access-kgqw5\") on node \"crc\" DevicePath \"\"" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.735845 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/434fbbdb-10c8-461e-812f-7725b27b0111-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "434fbbdb-10c8-461e-812f-7725b27b0111" (UID: "434fbbdb-10c8-461e-812f-7725b27b0111"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.743844 4739 scope.go:117] "RemoveContainer" containerID="d707e6fd0862a55ef17b0c0dda043f2301af28672b80c8c03e1fd2921f9114b3" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.765765 4739 scope.go:117] "RemoveContainer" containerID="60b7ed89dcd28ad8ff68122ff3a20ac1f5bebaaf5345959c8d57aab735b16d6f" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.802279 4739 scope.go:117] "RemoveContainer" containerID="dedcf54eafbcff7957421b1569934a5e4a9c7137a06b70c8c026ebfd4e63a614" Dec 01 15:58:57 crc kubenswrapper[4739]: E1201 15:58:57.804395 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dedcf54eafbcff7957421b1569934a5e4a9c7137a06b70c8c026ebfd4e63a614\": container with ID starting with dedcf54eafbcff7957421b1569934a5e4a9c7137a06b70c8c026ebfd4e63a614 not found: ID does not exist" containerID="dedcf54eafbcff7957421b1569934a5e4a9c7137a06b70c8c026ebfd4e63a614" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.804459 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dedcf54eafbcff7957421b1569934a5e4a9c7137a06b70c8c026ebfd4e63a614"} err="failed to get container status \"dedcf54eafbcff7957421b1569934a5e4a9c7137a06b70c8c026ebfd4e63a614\": rpc error: code = NotFound desc = could not find container \"dedcf54eafbcff7957421b1569934a5e4a9c7137a06b70c8c026ebfd4e63a614\": container with ID starting with dedcf54eafbcff7957421b1569934a5e4a9c7137a06b70c8c026ebfd4e63a614 not found: ID does not exist" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.804494 4739 scope.go:117] "RemoveContainer" containerID="d707e6fd0862a55ef17b0c0dda043f2301af28672b80c8c03e1fd2921f9114b3" Dec 01 15:58:57 crc kubenswrapper[4739]: E1201 15:58:57.804876 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d707e6fd0862a55ef17b0c0dda043f2301af28672b80c8c03e1fd2921f9114b3\": container with ID starting with d707e6fd0862a55ef17b0c0dda043f2301af28672b80c8c03e1fd2921f9114b3 not found: ID does not exist" containerID="d707e6fd0862a55ef17b0c0dda043f2301af28672b80c8c03e1fd2921f9114b3" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.804915 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d707e6fd0862a55ef17b0c0dda043f2301af28672b80c8c03e1fd2921f9114b3"} err="failed to get container status \"d707e6fd0862a55ef17b0c0dda043f2301af28672b80c8c03e1fd2921f9114b3\": rpc error: code = NotFound desc = could not find container \"d707e6fd0862a55ef17b0c0dda043f2301af28672b80c8c03e1fd2921f9114b3\": container with ID starting with d707e6fd0862a55ef17b0c0dda043f2301af28672b80c8c03e1fd2921f9114b3 not found: ID does not exist" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.804941 4739 scope.go:117] "RemoveContainer" containerID="60b7ed89dcd28ad8ff68122ff3a20ac1f5bebaaf5345959c8d57aab735b16d6f" Dec 01 15:58:57 crc kubenswrapper[4739]: E1201 15:58:57.805219 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60b7ed89dcd28ad8ff68122ff3a20ac1f5bebaaf5345959c8d57aab735b16d6f\": container with ID starting with 60b7ed89dcd28ad8ff68122ff3a20ac1f5bebaaf5345959c8d57aab735b16d6f not found: ID does not exist" containerID="60b7ed89dcd28ad8ff68122ff3a20ac1f5bebaaf5345959c8d57aab735b16d6f" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.805252 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60b7ed89dcd28ad8ff68122ff3a20ac1f5bebaaf5345959c8d57aab735b16d6f"} err="failed to get container status \"60b7ed89dcd28ad8ff68122ff3a20ac1f5bebaaf5345959c8d57aab735b16d6f\": rpc error: code = NotFound desc = could not find container \"60b7ed89dcd28ad8ff68122ff3a20ac1f5bebaaf5345959c8d57aab735b16d6f\": container with ID starting with 60b7ed89dcd28ad8ff68122ff3a20ac1f5bebaaf5345959c8d57aab735b16d6f not found: ID does not exist" Dec 01 15:58:57 crc kubenswrapper[4739]: I1201 15:58:57.829547 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434fbbdb-10c8-461e-812f-7725b27b0111-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:58:58 crc kubenswrapper[4739]: I1201 15:58:58.048952 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fnwbc"] Dec 01 15:58:58 crc kubenswrapper[4739]: I1201 15:58:58.056345 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fnwbc"] Dec 01 15:58:58 crc kubenswrapper[4739]: I1201 15:58:58.488172 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="434fbbdb-10c8-461e-812f-7725b27b0111" path="/var/lib/kubelet/pods/434fbbdb-10c8-461e-812f-7725b27b0111/volumes" Dec 01 15:59:09 crc kubenswrapper[4739]: I1201 15:59:09.622005 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:59:09 crc kubenswrapper[4739]: I1201 15:59:09.624593 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:59:20 crc kubenswrapper[4739]: I1201 15:59:20.468667 4739 scope.go:117] "RemoveContainer" containerID="9665bb75e988ad1593824068baacfbd407afe31d88d02b82b9fbe975a9a2329d" Dec 01 15:59:39 crc kubenswrapper[4739]: I1201 15:59:39.625061 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:59:39 crc kubenswrapper[4739]: I1201 15:59:39.626486 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:59:39 crc kubenswrapper[4739]: I1201 15:59:39.626641 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 15:59:39 crc kubenswrapper[4739]: I1201 15:59:39.627561 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:59:39 crc kubenswrapper[4739]: I1201 15:59:39.627694 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" gracePeriod=600 Dec 01 15:59:39 crc kubenswrapper[4739]: E1201 15:59:39.752739 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 15:59:40 crc kubenswrapper[4739]: I1201 15:59:40.246106 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" exitCode=0 Dec 01 15:59:40 crc kubenswrapper[4739]: I1201 15:59:40.246173 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a"} Dec 01 15:59:40 crc kubenswrapper[4739]: I1201 15:59:40.246232 4739 scope.go:117] "RemoveContainer" containerID="faad8f9bb664920c3ea11c12b9b5f870015e0ec81598417117215899a38ccfdb" Dec 01 15:59:40 crc kubenswrapper[4739]: I1201 15:59:40.246871 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 15:59:40 crc kubenswrapper[4739]: E1201 15:59:40.247239 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 15:59:51 crc kubenswrapper[4739]: I1201 15:59:51.324929 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 15:59:51 crc kubenswrapper[4739]: E1201 15:59:51.333008 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.166198 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq"] Dec 01 16:00:00 crc kubenswrapper[4739]: E1201 16:00:00.167383 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434fbbdb-10c8-461e-812f-7725b27b0111" containerName="extract-content" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.167403 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="434fbbdb-10c8-461e-812f-7725b27b0111" containerName="extract-content" Dec 01 16:00:00 crc kubenswrapper[4739]: E1201 16:00:00.167628 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434fbbdb-10c8-461e-812f-7725b27b0111" containerName="registry-server" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.167652 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="434fbbdb-10c8-461e-812f-7725b27b0111" containerName="registry-server" Dec 01 16:00:00 crc kubenswrapper[4739]: E1201 16:00:00.167674 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434fbbdb-10c8-461e-812f-7725b27b0111" containerName="extract-utilities" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.167687 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="434fbbdb-10c8-461e-812f-7725b27b0111" containerName="extract-utilities" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.167978 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="434fbbdb-10c8-461e-812f-7725b27b0111" containerName="registry-server" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.168871 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.173000 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.173086 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.204468 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq"] Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.280897 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/74b76c6e-6b96-4361-b887-8b3d3779235c-secret-volume\") pod \"collect-profiles-29410080-7tttq\" (UID: \"74b76c6e-6b96-4361-b887-8b3d3779235c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.280982 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgbkz\" (UniqueName: \"kubernetes.io/projected/74b76c6e-6b96-4361-b887-8b3d3779235c-kube-api-access-wgbkz\") pod \"collect-profiles-29410080-7tttq\" (UID: \"74b76c6e-6b96-4361-b887-8b3d3779235c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.281048 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/74b76c6e-6b96-4361-b887-8b3d3779235c-config-volume\") pod \"collect-profiles-29410080-7tttq\" (UID: \"74b76c6e-6b96-4361-b887-8b3d3779235c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.382674 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/74b76c6e-6b96-4361-b887-8b3d3779235c-config-volume\") pod \"collect-profiles-29410080-7tttq\" (UID: \"74b76c6e-6b96-4361-b887-8b3d3779235c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.382795 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/74b76c6e-6b96-4361-b887-8b3d3779235c-secret-volume\") pod \"collect-profiles-29410080-7tttq\" (UID: \"74b76c6e-6b96-4361-b887-8b3d3779235c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.382832 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgbkz\" (UniqueName: \"kubernetes.io/projected/74b76c6e-6b96-4361-b887-8b3d3779235c-kube-api-access-wgbkz\") pod \"collect-profiles-29410080-7tttq\" (UID: \"74b76c6e-6b96-4361-b887-8b3d3779235c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.383524 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/74b76c6e-6b96-4361-b887-8b3d3779235c-config-volume\") pod \"collect-profiles-29410080-7tttq\" (UID: \"74b76c6e-6b96-4361-b887-8b3d3779235c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.389661 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/74b76c6e-6b96-4361-b887-8b3d3779235c-secret-volume\") pod \"collect-profiles-29410080-7tttq\" (UID: \"74b76c6e-6b96-4361-b887-8b3d3779235c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.404366 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgbkz\" (UniqueName: \"kubernetes.io/projected/74b76c6e-6b96-4361-b887-8b3d3779235c-kube-api-access-wgbkz\") pod \"collect-profiles-29410080-7tttq\" (UID: \"74b76c6e-6b96-4361-b887-8b3d3779235c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.494907 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" Dec 01 16:00:00 crc kubenswrapper[4739]: I1201 16:00:00.914359 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq"] Dec 01 16:00:00 crc kubenswrapper[4739]: W1201 16:00:00.927467 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74b76c6e_6b96_4361_b887_8b3d3779235c.slice/crio-1ede56c058d79055a2a5b85966e00a4c85cdb745eeefa1579756fc1bc4e2b0d7 WatchSource:0}: Error finding container 1ede56c058d79055a2a5b85966e00a4c85cdb745eeefa1579756fc1bc4e2b0d7: Status 404 returned error can't find the container with id 1ede56c058d79055a2a5b85966e00a4c85cdb745eeefa1579756fc1bc4e2b0d7 Dec 01 16:00:01 crc kubenswrapper[4739]: I1201 16:00:01.493133 4739 generic.go:334] "Generic (PLEG): container finished" podID="74b76c6e-6b96-4361-b887-8b3d3779235c" containerID="033a74d0d0712e84a4423c97a7a17ee50087258363dec676c9a8a2f68c4cb433" exitCode=0 Dec 01 16:00:01 crc kubenswrapper[4739]: I1201 16:00:01.494260 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" event={"ID":"74b76c6e-6b96-4361-b887-8b3d3779235c","Type":"ContainerDied","Data":"033a74d0d0712e84a4423c97a7a17ee50087258363dec676c9a8a2f68c4cb433"} Dec 01 16:00:01 crc kubenswrapper[4739]: I1201 16:00:01.494336 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" event={"ID":"74b76c6e-6b96-4361-b887-8b3d3779235c","Type":"ContainerStarted","Data":"1ede56c058d79055a2a5b85966e00a4c85cdb745eeefa1579756fc1bc4e2b0d7"} Dec 01 16:00:02 crc kubenswrapper[4739]: I1201 16:00:02.924500 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" Dec 01 16:00:03 crc kubenswrapper[4739]: I1201 16:00:03.033384 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/74b76c6e-6b96-4361-b887-8b3d3779235c-config-volume\") pod \"74b76c6e-6b96-4361-b887-8b3d3779235c\" (UID: \"74b76c6e-6b96-4361-b887-8b3d3779235c\") " Dec 01 16:00:03 crc kubenswrapper[4739]: I1201 16:00:03.033542 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgbkz\" (UniqueName: \"kubernetes.io/projected/74b76c6e-6b96-4361-b887-8b3d3779235c-kube-api-access-wgbkz\") pod \"74b76c6e-6b96-4361-b887-8b3d3779235c\" (UID: \"74b76c6e-6b96-4361-b887-8b3d3779235c\") " Dec 01 16:00:03 crc kubenswrapper[4739]: I1201 16:00:03.033619 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/74b76c6e-6b96-4361-b887-8b3d3779235c-secret-volume\") pod \"74b76c6e-6b96-4361-b887-8b3d3779235c\" (UID: \"74b76c6e-6b96-4361-b887-8b3d3779235c\") " Dec 01 16:00:03 crc kubenswrapper[4739]: I1201 16:00:03.034537 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74b76c6e-6b96-4361-b887-8b3d3779235c-config-volume" (OuterVolumeSpecName: "config-volume") pod "74b76c6e-6b96-4361-b887-8b3d3779235c" (UID: "74b76c6e-6b96-4361-b887-8b3d3779235c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:00:03 crc kubenswrapper[4739]: I1201 16:00:03.039346 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74b76c6e-6b96-4361-b887-8b3d3779235c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "74b76c6e-6b96-4361-b887-8b3d3779235c" (UID: "74b76c6e-6b96-4361-b887-8b3d3779235c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:00:03 crc kubenswrapper[4739]: I1201 16:00:03.040180 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74b76c6e-6b96-4361-b887-8b3d3779235c-kube-api-access-wgbkz" (OuterVolumeSpecName: "kube-api-access-wgbkz") pod "74b76c6e-6b96-4361-b887-8b3d3779235c" (UID: "74b76c6e-6b96-4361-b887-8b3d3779235c"). InnerVolumeSpecName "kube-api-access-wgbkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:00:03 crc kubenswrapper[4739]: I1201 16:00:03.135708 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgbkz\" (UniqueName: \"kubernetes.io/projected/74b76c6e-6b96-4361-b887-8b3d3779235c-kube-api-access-wgbkz\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:03 crc kubenswrapper[4739]: I1201 16:00:03.135753 4739 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/74b76c6e-6b96-4361-b887-8b3d3779235c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:03 crc kubenswrapper[4739]: I1201 16:00:03.135767 4739 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/74b76c6e-6b96-4361-b887-8b3d3779235c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:03 crc kubenswrapper[4739]: I1201 16:00:03.517398 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" event={"ID":"74b76c6e-6b96-4361-b887-8b3d3779235c","Type":"ContainerDied","Data":"1ede56c058d79055a2a5b85966e00a4c85cdb745eeefa1579756fc1bc4e2b0d7"} Dec 01 16:00:03 crc kubenswrapper[4739]: I1201 16:00:03.517464 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ede56c058d79055a2a5b85966e00a4c85cdb745eeefa1579756fc1bc4e2b0d7" Dec 01 16:00:03 crc kubenswrapper[4739]: I1201 16:00:03.517509 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq" Dec 01 16:00:04 crc kubenswrapper[4739]: I1201 16:00:04.850888 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w7kg7"] Dec 01 16:00:04 crc kubenswrapper[4739]: E1201 16:00:04.851729 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b76c6e-6b96-4361-b887-8b3d3779235c" containerName="collect-profiles" Dec 01 16:00:04 crc kubenswrapper[4739]: I1201 16:00:04.851746 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b76c6e-6b96-4361-b887-8b3d3779235c" containerName="collect-profiles" Dec 01 16:00:04 crc kubenswrapper[4739]: I1201 16:00:04.851948 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="74b76c6e-6b96-4361-b887-8b3d3779235c" containerName="collect-profiles" Dec 01 16:00:04 crc kubenswrapper[4739]: I1201 16:00:04.853766 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:04 crc kubenswrapper[4739]: I1201 16:00:04.871547 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w7kg7"] Dec 01 16:00:04 crc kubenswrapper[4739]: I1201 16:00:04.974775 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-utilities\") pod \"community-operators-w7kg7\" (UID: \"50d98cc2-b2b9-40f1-ac80-28585dceb6b9\") " pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:04 crc kubenswrapper[4739]: I1201 16:00:04.974855 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5v6h\" (UniqueName: \"kubernetes.io/projected/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-kube-api-access-m5v6h\") pod \"community-operators-w7kg7\" (UID: \"50d98cc2-b2b9-40f1-ac80-28585dceb6b9\") " pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:04 crc kubenswrapper[4739]: I1201 16:00:04.975242 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-catalog-content\") pod \"community-operators-w7kg7\" (UID: \"50d98cc2-b2b9-40f1-ac80-28585dceb6b9\") " pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:05 crc kubenswrapper[4739]: I1201 16:00:05.077285 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5v6h\" (UniqueName: \"kubernetes.io/projected/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-kube-api-access-m5v6h\") pod \"community-operators-w7kg7\" (UID: \"50d98cc2-b2b9-40f1-ac80-28585dceb6b9\") " pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:05 crc kubenswrapper[4739]: I1201 16:00:05.077669 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-catalog-content\") pod \"community-operators-w7kg7\" (UID: \"50d98cc2-b2b9-40f1-ac80-28585dceb6b9\") " pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:05 crc kubenswrapper[4739]: I1201 16:00:05.077982 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-utilities\") pod \"community-operators-w7kg7\" (UID: \"50d98cc2-b2b9-40f1-ac80-28585dceb6b9\") " pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:05 crc kubenswrapper[4739]: I1201 16:00:05.078063 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-catalog-content\") pod \"community-operators-w7kg7\" (UID: \"50d98cc2-b2b9-40f1-ac80-28585dceb6b9\") " pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:05 crc kubenswrapper[4739]: I1201 16:00:05.078409 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-utilities\") pod \"community-operators-w7kg7\" (UID: \"50d98cc2-b2b9-40f1-ac80-28585dceb6b9\") " pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:05 crc kubenswrapper[4739]: I1201 16:00:05.099682 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5v6h\" (UniqueName: \"kubernetes.io/projected/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-kube-api-access-m5v6h\") pod \"community-operators-w7kg7\" (UID: \"50d98cc2-b2b9-40f1-ac80-28585dceb6b9\") " pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:05 crc kubenswrapper[4739]: I1201 16:00:05.195099 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:05 crc kubenswrapper[4739]: I1201 16:00:05.697626 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w7kg7"] Dec 01 16:00:05 crc kubenswrapper[4739]: W1201 16:00:05.698776 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50d98cc2_b2b9_40f1_ac80_28585dceb6b9.slice/crio-25db81957021c48135725e0069b7ebf09a7960c58eff7b9379f86fa12c8c340e WatchSource:0}: Error finding container 25db81957021c48135725e0069b7ebf09a7960c58eff7b9379f86fa12c8c340e: Status 404 returned error can't find the container with id 25db81957021c48135725e0069b7ebf09a7960c58eff7b9379f86fa12c8c340e Dec 01 16:00:06 crc kubenswrapper[4739]: I1201 16:00:06.478211 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:00:06 crc kubenswrapper[4739]: E1201 16:00:06.479082 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:00:06 crc kubenswrapper[4739]: I1201 16:00:06.544660 4739 generic.go:334] "Generic (PLEG): container finished" podID="50d98cc2-b2b9-40f1-ac80-28585dceb6b9" containerID="a2aa12cff190e8438c1b3469cbcb3ba18774e2b578f234079616a63802dc0a89" exitCode=0 Dec 01 16:00:06 crc kubenswrapper[4739]: I1201 16:00:06.544706 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w7kg7" event={"ID":"50d98cc2-b2b9-40f1-ac80-28585dceb6b9","Type":"ContainerDied","Data":"a2aa12cff190e8438c1b3469cbcb3ba18774e2b578f234079616a63802dc0a89"} Dec 01 16:00:06 crc kubenswrapper[4739]: I1201 16:00:06.544734 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w7kg7" event={"ID":"50d98cc2-b2b9-40f1-ac80-28585dceb6b9","Type":"ContainerStarted","Data":"25db81957021c48135725e0069b7ebf09a7960c58eff7b9379f86fa12c8c340e"} Dec 01 16:00:08 crc kubenswrapper[4739]: I1201 16:00:08.574023 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w7kg7" event={"ID":"50d98cc2-b2b9-40f1-ac80-28585dceb6b9","Type":"ContainerStarted","Data":"4f0f98efea3660897a1eb31c33bab69f4ea7b0d83c31873e48e21ecd683af132"} Dec 01 16:00:09 crc kubenswrapper[4739]: I1201 16:00:09.590504 4739 generic.go:334] "Generic (PLEG): container finished" podID="50d98cc2-b2b9-40f1-ac80-28585dceb6b9" containerID="4f0f98efea3660897a1eb31c33bab69f4ea7b0d83c31873e48e21ecd683af132" exitCode=0 Dec 01 16:00:09 crc kubenswrapper[4739]: I1201 16:00:09.590752 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w7kg7" event={"ID":"50d98cc2-b2b9-40f1-ac80-28585dceb6b9","Type":"ContainerDied","Data":"4f0f98efea3660897a1eb31c33bab69f4ea7b0d83c31873e48e21ecd683af132"} Dec 01 16:00:10 crc kubenswrapper[4739]: I1201 16:00:10.603539 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w7kg7" event={"ID":"50d98cc2-b2b9-40f1-ac80-28585dceb6b9","Type":"ContainerStarted","Data":"ec8d76a78ff2dc262bed37dcbd921cb8b63fbe85695b4fd6ad070bd07838d7e9"} Dec 01 16:00:10 crc kubenswrapper[4739]: I1201 16:00:10.634998 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w7kg7" podStartSLOduration=3.048355472 podStartE2EDuration="6.634979033s" podCreationTimestamp="2025-12-01 16:00:04 +0000 UTC" firstStartedPulling="2025-12-01 16:00:06.546649952 +0000 UTC m=+1508.372396046" lastFinishedPulling="2025-12-01 16:00:10.133273503 +0000 UTC m=+1511.959019607" observedRunningTime="2025-12-01 16:00:10.62584131 +0000 UTC m=+1512.451587404" watchObservedRunningTime="2025-12-01 16:00:10.634979033 +0000 UTC m=+1512.460725127" Dec 01 16:00:15 crc kubenswrapper[4739]: I1201 16:00:15.195997 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:15 crc kubenswrapper[4739]: I1201 16:00:15.196652 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:15 crc kubenswrapper[4739]: I1201 16:00:15.289196 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:15 crc kubenswrapper[4739]: I1201 16:00:15.717170 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:15 crc kubenswrapper[4739]: I1201 16:00:15.776068 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w7kg7"] Dec 01 16:00:17 crc kubenswrapper[4739]: I1201 16:00:17.671832 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w7kg7" podUID="50d98cc2-b2b9-40f1-ac80-28585dceb6b9" containerName="registry-server" containerID="cri-o://ec8d76a78ff2dc262bed37dcbd921cb8b63fbe85695b4fd6ad070bd07838d7e9" gracePeriod=2 Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.187961 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.352461 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-catalog-content\") pod \"50d98cc2-b2b9-40f1-ac80-28585dceb6b9\" (UID: \"50d98cc2-b2b9-40f1-ac80-28585dceb6b9\") " Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.352668 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5v6h\" (UniqueName: \"kubernetes.io/projected/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-kube-api-access-m5v6h\") pod \"50d98cc2-b2b9-40f1-ac80-28585dceb6b9\" (UID: \"50d98cc2-b2b9-40f1-ac80-28585dceb6b9\") " Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.352730 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-utilities\") pod \"50d98cc2-b2b9-40f1-ac80-28585dceb6b9\" (UID: \"50d98cc2-b2b9-40f1-ac80-28585dceb6b9\") " Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.354205 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-utilities" (OuterVolumeSpecName: "utilities") pod "50d98cc2-b2b9-40f1-ac80-28585dceb6b9" (UID: "50d98cc2-b2b9-40f1-ac80-28585dceb6b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.359051 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-kube-api-access-m5v6h" (OuterVolumeSpecName: "kube-api-access-m5v6h") pod "50d98cc2-b2b9-40f1-ac80-28585dceb6b9" (UID: "50d98cc2-b2b9-40f1-ac80-28585dceb6b9"). InnerVolumeSpecName "kube-api-access-m5v6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.435299 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50d98cc2-b2b9-40f1-ac80-28585dceb6b9" (UID: "50d98cc2-b2b9-40f1-ac80-28585dceb6b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.456291 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5v6h\" (UniqueName: \"kubernetes.io/projected/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-kube-api-access-m5v6h\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.456337 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.456356 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50d98cc2-b2b9-40f1-ac80-28585dceb6b9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.686710 4739 generic.go:334] "Generic (PLEG): container finished" podID="50d98cc2-b2b9-40f1-ac80-28585dceb6b9" containerID="ec8d76a78ff2dc262bed37dcbd921cb8b63fbe85695b4fd6ad070bd07838d7e9" exitCode=0 Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.686787 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w7kg7" event={"ID":"50d98cc2-b2b9-40f1-ac80-28585dceb6b9","Type":"ContainerDied","Data":"ec8d76a78ff2dc262bed37dcbd921cb8b63fbe85695b4fd6ad070bd07838d7e9"} Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.686844 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w7kg7" event={"ID":"50d98cc2-b2b9-40f1-ac80-28585dceb6b9","Type":"ContainerDied","Data":"25db81957021c48135725e0069b7ebf09a7960c58eff7b9379f86fa12c8c340e"} Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.686847 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w7kg7" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.686879 4739 scope.go:117] "RemoveContainer" containerID="ec8d76a78ff2dc262bed37dcbd921cb8b63fbe85695b4fd6ad070bd07838d7e9" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.725476 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w7kg7"] Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.740183 4739 scope.go:117] "RemoveContainer" containerID="4f0f98efea3660897a1eb31c33bab69f4ea7b0d83c31873e48e21ecd683af132" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.742296 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w7kg7"] Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.767941 4739 scope.go:117] "RemoveContainer" containerID="a2aa12cff190e8438c1b3469cbcb3ba18774e2b578f234079616a63802dc0a89" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.803929 4739 scope.go:117] "RemoveContainer" containerID="ec8d76a78ff2dc262bed37dcbd921cb8b63fbe85695b4fd6ad070bd07838d7e9" Dec 01 16:00:18 crc kubenswrapper[4739]: E1201 16:00:18.804492 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec8d76a78ff2dc262bed37dcbd921cb8b63fbe85695b4fd6ad070bd07838d7e9\": container with ID starting with ec8d76a78ff2dc262bed37dcbd921cb8b63fbe85695b4fd6ad070bd07838d7e9 not found: ID does not exist" containerID="ec8d76a78ff2dc262bed37dcbd921cb8b63fbe85695b4fd6ad070bd07838d7e9" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.804530 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec8d76a78ff2dc262bed37dcbd921cb8b63fbe85695b4fd6ad070bd07838d7e9"} err="failed to get container status \"ec8d76a78ff2dc262bed37dcbd921cb8b63fbe85695b4fd6ad070bd07838d7e9\": rpc error: code = NotFound desc = could not find container \"ec8d76a78ff2dc262bed37dcbd921cb8b63fbe85695b4fd6ad070bd07838d7e9\": container with ID starting with ec8d76a78ff2dc262bed37dcbd921cb8b63fbe85695b4fd6ad070bd07838d7e9 not found: ID does not exist" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.804556 4739 scope.go:117] "RemoveContainer" containerID="4f0f98efea3660897a1eb31c33bab69f4ea7b0d83c31873e48e21ecd683af132" Dec 01 16:00:18 crc kubenswrapper[4739]: E1201 16:00:18.805057 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f0f98efea3660897a1eb31c33bab69f4ea7b0d83c31873e48e21ecd683af132\": container with ID starting with 4f0f98efea3660897a1eb31c33bab69f4ea7b0d83c31873e48e21ecd683af132 not found: ID does not exist" containerID="4f0f98efea3660897a1eb31c33bab69f4ea7b0d83c31873e48e21ecd683af132" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.805087 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f0f98efea3660897a1eb31c33bab69f4ea7b0d83c31873e48e21ecd683af132"} err="failed to get container status \"4f0f98efea3660897a1eb31c33bab69f4ea7b0d83c31873e48e21ecd683af132\": rpc error: code = NotFound desc = could not find container \"4f0f98efea3660897a1eb31c33bab69f4ea7b0d83c31873e48e21ecd683af132\": container with ID starting with 4f0f98efea3660897a1eb31c33bab69f4ea7b0d83c31873e48e21ecd683af132 not found: ID does not exist" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.805105 4739 scope.go:117] "RemoveContainer" containerID="a2aa12cff190e8438c1b3469cbcb3ba18774e2b578f234079616a63802dc0a89" Dec 01 16:00:18 crc kubenswrapper[4739]: E1201 16:00:18.805673 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2aa12cff190e8438c1b3469cbcb3ba18774e2b578f234079616a63802dc0a89\": container with ID starting with a2aa12cff190e8438c1b3469cbcb3ba18774e2b578f234079616a63802dc0a89 not found: ID does not exist" containerID="a2aa12cff190e8438c1b3469cbcb3ba18774e2b578f234079616a63802dc0a89" Dec 01 16:00:18 crc kubenswrapper[4739]: I1201 16:00:18.805700 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2aa12cff190e8438c1b3469cbcb3ba18774e2b578f234079616a63802dc0a89"} err="failed to get container status \"a2aa12cff190e8438c1b3469cbcb3ba18774e2b578f234079616a63802dc0a89\": rpc error: code = NotFound desc = could not find container \"a2aa12cff190e8438c1b3469cbcb3ba18774e2b578f234079616a63802dc0a89\": container with ID starting with a2aa12cff190e8438c1b3469cbcb3ba18774e2b578f234079616a63802dc0a89 not found: ID does not exist" Dec 01 16:00:20 crc kubenswrapper[4739]: I1201 16:00:20.487915 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50d98cc2-b2b9-40f1-ac80-28585dceb6b9" path="/var/lib/kubelet/pods/50d98cc2-b2b9-40f1-ac80-28585dceb6b9/volumes" Dec 01 16:00:20 crc kubenswrapper[4739]: I1201 16:00:20.598211 4739 scope.go:117] "RemoveContainer" containerID="ca3820b86fb7dfa5f667c711afb764f9439cd7ced3610355bed4b0e880bdcfd1" Dec 01 16:00:20 crc kubenswrapper[4739]: I1201 16:00:20.630153 4739 scope.go:117] "RemoveContainer" containerID="be779c8cac2cdd9fa5ea82a05062e9221ae3999b7c4fbfe6b4e72cedfe35a461" Dec 01 16:00:20 crc kubenswrapper[4739]: I1201 16:00:20.655553 4739 scope.go:117] "RemoveContainer" containerID="0cc96695e1e1284653477ead72552d8628ee825acd11d60f11e0811b31143f55" Dec 01 16:00:20 crc kubenswrapper[4739]: I1201 16:00:20.670332 4739 scope.go:117] "RemoveContainer" containerID="d64c02087f02dbbb8591dee821258b2698108bde07bfdd6549346b4407cd56f8" Dec 01 16:00:21 crc kubenswrapper[4739]: I1201 16:00:21.477885 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:00:21 crc kubenswrapper[4739]: E1201 16:00:21.478230 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:00:24 crc kubenswrapper[4739]: I1201 16:00:24.754701 4739 generic.go:334] "Generic (PLEG): container finished" podID="ae86132a-1c2e-4e6b-9b3a-6f8b63be4270" containerID="cb26e7e043112190326b4e04beaaab20e5f5d7b71e3274844c6b90e1c92d063b" exitCode=0 Dec 01 16:00:24 crc kubenswrapper[4739]: I1201 16:00:24.754833 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" event={"ID":"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270","Type":"ContainerDied","Data":"cb26e7e043112190326b4e04beaaab20e5f5d7b71e3274844c6b90e1c92d063b"} Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.251088 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.410508 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-inventory\") pod \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.411840 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-bootstrap-combined-ca-bundle\") pod \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.411880 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-ssh-key\") pod \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.411980 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2krp2\" (UniqueName: \"kubernetes.io/projected/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-kube-api-access-2krp2\") pod \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\" (UID: \"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270\") " Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.419236 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-kube-api-access-2krp2" (OuterVolumeSpecName: "kube-api-access-2krp2") pod "ae86132a-1c2e-4e6b-9b3a-6f8b63be4270" (UID: "ae86132a-1c2e-4e6b-9b3a-6f8b63be4270"). InnerVolumeSpecName "kube-api-access-2krp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.420488 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "ae86132a-1c2e-4e6b-9b3a-6f8b63be4270" (UID: "ae86132a-1c2e-4e6b-9b3a-6f8b63be4270"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.445197 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ae86132a-1c2e-4e6b-9b3a-6f8b63be4270" (UID: "ae86132a-1c2e-4e6b-9b3a-6f8b63be4270"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.464440 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-inventory" (OuterVolumeSpecName: "inventory") pod "ae86132a-1c2e-4e6b-9b3a-6f8b63be4270" (UID: "ae86132a-1c2e-4e6b-9b3a-6f8b63be4270"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.517160 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.517208 4739 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.517221 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.517231 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2krp2\" (UniqueName: \"kubernetes.io/projected/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270-kube-api-access-2krp2\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.781490 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" event={"ID":"ae86132a-1c2e-4e6b-9b3a-6f8b63be4270","Type":"ContainerDied","Data":"5a602b5c47a3a35d80a9ec3eb576f31799f83e27015048af199421eea241b4e4"} Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.781557 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a602b5c47a3a35d80a9ec3eb576f31799f83e27015048af199421eea241b4e4" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.781699 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.897654 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4"] Dec 01 16:00:26 crc kubenswrapper[4739]: E1201 16:00:26.898581 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50d98cc2-b2b9-40f1-ac80-28585dceb6b9" containerName="registry-server" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.898621 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="50d98cc2-b2b9-40f1-ac80-28585dceb6b9" containerName="registry-server" Dec 01 16:00:26 crc kubenswrapper[4739]: E1201 16:00:26.898660 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae86132a-1c2e-4e6b-9b3a-6f8b63be4270" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.898684 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae86132a-1c2e-4e6b-9b3a-6f8b63be4270" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 16:00:26 crc kubenswrapper[4739]: E1201 16:00:26.898750 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50d98cc2-b2b9-40f1-ac80-28585dceb6b9" containerName="extract-utilities" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.898764 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="50d98cc2-b2b9-40f1-ac80-28585dceb6b9" containerName="extract-utilities" Dec 01 16:00:26 crc kubenswrapper[4739]: E1201 16:00:26.898785 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50d98cc2-b2b9-40f1-ac80-28585dceb6b9" containerName="extract-content" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.898798 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="50d98cc2-b2b9-40f1-ac80-28585dceb6b9" containerName="extract-content" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.899173 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae86132a-1c2e-4e6b-9b3a-6f8b63be4270" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.899202 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="50d98cc2-b2b9-40f1-ac80-28585dceb6b9" containerName="registry-server" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.900387 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.903801 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.904692 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.904793 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.904794 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:00:26 crc kubenswrapper[4739]: I1201 16:00:26.909319 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4"] Dec 01 16:00:27 crc kubenswrapper[4739]: I1201 16:00:27.027344 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c7d3cbb-19d3-4855-8ad4-72504ba66787-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4\" (UID: \"5c7d3cbb-19d3-4855-8ad4-72504ba66787\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" Dec 01 16:00:27 crc kubenswrapper[4739]: I1201 16:00:27.027873 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86msw\" (UniqueName: \"kubernetes.io/projected/5c7d3cbb-19d3-4855-8ad4-72504ba66787-kube-api-access-86msw\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4\" (UID: \"5c7d3cbb-19d3-4855-8ad4-72504ba66787\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" Dec 01 16:00:27 crc kubenswrapper[4739]: I1201 16:00:27.027977 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c7d3cbb-19d3-4855-8ad4-72504ba66787-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4\" (UID: \"5c7d3cbb-19d3-4855-8ad4-72504ba66787\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" Dec 01 16:00:27 crc kubenswrapper[4739]: I1201 16:00:27.130284 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c7d3cbb-19d3-4855-8ad4-72504ba66787-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4\" (UID: \"5c7d3cbb-19d3-4855-8ad4-72504ba66787\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" Dec 01 16:00:27 crc kubenswrapper[4739]: I1201 16:00:27.130609 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86msw\" (UniqueName: \"kubernetes.io/projected/5c7d3cbb-19d3-4855-8ad4-72504ba66787-kube-api-access-86msw\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4\" (UID: \"5c7d3cbb-19d3-4855-8ad4-72504ba66787\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" Dec 01 16:00:27 crc kubenswrapper[4739]: I1201 16:00:27.130682 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c7d3cbb-19d3-4855-8ad4-72504ba66787-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4\" (UID: \"5c7d3cbb-19d3-4855-8ad4-72504ba66787\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" Dec 01 16:00:27 crc kubenswrapper[4739]: I1201 16:00:27.135613 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c7d3cbb-19d3-4855-8ad4-72504ba66787-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4\" (UID: \"5c7d3cbb-19d3-4855-8ad4-72504ba66787\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" Dec 01 16:00:27 crc kubenswrapper[4739]: I1201 16:00:27.137214 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c7d3cbb-19d3-4855-8ad4-72504ba66787-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4\" (UID: \"5c7d3cbb-19d3-4855-8ad4-72504ba66787\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" Dec 01 16:00:27 crc kubenswrapper[4739]: I1201 16:00:27.147247 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86msw\" (UniqueName: \"kubernetes.io/projected/5c7d3cbb-19d3-4855-8ad4-72504ba66787-kube-api-access-86msw\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4\" (UID: \"5c7d3cbb-19d3-4855-8ad4-72504ba66787\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" Dec 01 16:00:27 crc kubenswrapper[4739]: I1201 16:00:27.231373 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" Dec 01 16:00:27 crc kubenswrapper[4739]: I1201 16:00:27.917876 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4"] Dec 01 16:00:28 crc kubenswrapper[4739]: I1201 16:00:28.820034 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" event={"ID":"5c7d3cbb-19d3-4855-8ad4-72504ba66787","Type":"ContainerStarted","Data":"a9c970c851d4a98da0388150708d0b25d357f5e7c7e769e7c67f720160e381e3"} Dec 01 16:00:29 crc kubenswrapper[4739]: I1201 16:00:29.831563 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" event={"ID":"5c7d3cbb-19d3-4855-8ad4-72504ba66787","Type":"ContainerStarted","Data":"3fac8a90bed2abfaa40fef7599e3db6a3a00c3706112ac25c5890105b83ed039"} Dec 01 16:00:29 crc kubenswrapper[4739]: I1201 16:00:29.855336 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" podStartSLOduration=3.028349565 podStartE2EDuration="3.855308063s" podCreationTimestamp="2025-12-01 16:00:26 +0000 UTC" firstStartedPulling="2025-12-01 16:00:27.928189391 +0000 UTC m=+1529.753935485" lastFinishedPulling="2025-12-01 16:00:28.755147859 +0000 UTC m=+1530.580893983" observedRunningTime="2025-12-01 16:00:29.852545568 +0000 UTC m=+1531.678291672" watchObservedRunningTime="2025-12-01 16:00:29.855308063 +0000 UTC m=+1531.681054157" Dec 01 16:00:32 crc kubenswrapper[4739]: I1201 16:00:32.478309 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:00:32 crc kubenswrapper[4739]: E1201 16:00:32.480910 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:00:45 crc kubenswrapper[4739]: I1201 16:00:45.477494 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:00:45 crc kubenswrapper[4739]: E1201 16:00:45.478975 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.159591 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29410081-vqlxx"] Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.161876 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.163497 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stf4q\" (UniqueName: \"kubernetes.io/projected/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-kube-api-access-stf4q\") pod \"keystone-cron-29410081-vqlxx\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.163568 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-fernet-keys\") pod \"keystone-cron-29410081-vqlxx\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.163616 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-combined-ca-bundle\") pod \"keystone-cron-29410081-vqlxx\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.163729 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-config-data\") pod \"keystone-cron-29410081-vqlxx\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.180961 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29410081-vqlxx"] Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.265894 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stf4q\" (UniqueName: \"kubernetes.io/projected/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-kube-api-access-stf4q\") pod \"keystone-cron-29410081-vqlxx\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.265934 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-fernet-keys\") pod \"keystone-cron-29410081-vqlxx\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.265977 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-combined-ca-bundle\") pod \"keystone-cron-29410081-vqlxx\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.266021 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-config-data\") pod \"keystone-cron-29410081-vqlxx\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.272556 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-config-data\") pod \"keystone-cron-29410081-vqlxx\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.272583 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-fernet-keys\") pod \"keystone-cron-29410081-vqlxx\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.278357 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-combined-ca-bundle\") pod \"keystone-cron-29410081-vqlxx\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.283099 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stf4q\" (UniqueName: \"kubernetes.io/projected/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-kube-api-access-stf4q\") pod \"keystone-cron-29410081-vqlxx\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.477581 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:01:00 crc kubenswrapper[4739]: E1201 16:01:00.478176 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.493806 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:00 crc kubenswrapper[4739]: I1201 16:01:00.943219 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29410081-vqlxx"] Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.162489 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410081-vqlxx" event={"ID":"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008","Type":"ContainerStarted","Data":"23ac4332746c5b9177f7f68b33b945ca3f78f459442d4ab4c5fa0894a74dae43"} Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.162533 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410081-vqlxx" event={"ID":"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008","Type":"ContainerStarted","Data":"b0cd99bf7d4fb7e93595c25fecbe68686c3e7530f121adc4e31e679ade4c33a5"} Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.184990 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29410081-vqlxx" podStartSLOduration=1.184966603 podStartE2EDuration="1.184966603s" podCreationTimestamp="2025-12-01 16:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:01:01.181325271 +0000 UTC m=+1563.007071375" watchObservedRunningTime="2025-12-01 16:01:01.184966603 +0000 UTC m=+1563.010712697" Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.618915 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w7nfj"] Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.621876 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.630602 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w7nfj"] Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.794284 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq8g9\" (UniqueName: \"kubernetes.io/projected/875e9c09-a5e1-455a-b12e-e46ea43be9ed-kube-api-access-tq8g9\") pod \"certified-operators-w7nfj\" (UID: \"875e9c09-a5e1-455a-b12e-e46ea43be9ed\") " pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.794395 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/875e9c09-a5e1-455a-b12e-e46ea43be9ed-catalog-content\") pod \"certified-operators-w7nfj\" (UID: \"875e9c09-a5e1-455a-b12e-e46ea43be9ed\") " pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.794454 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/875e9c09-a5e1-455a-b12e-e46ea43be9ed-utilities\") pod \"certified-operators-w7nfj\" (UID: \"875e9c09-a5e1-455a-b12e-e46ea43be9ed\") " pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.896663 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq8g9\" (UniqueName: \"kubernetes.io/projected/875e9c09-a5e1-455a-b12e-e46ea43be9ed-kube-api-access-tq8g9\") pod \"certified-operators-w7nfj\" (UID: \"875e9c09-a5e1-455a-b12e-e46ea43be9ed\") " pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.896729 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/875e9c09-a5e1-455a-b12e-e46ea43be9ed-catalog-content\") pod \"certified-operators-w7nfj\" (UID: \"875e9c09-a5e1-455a-b12e-e46ea43be9ed\") " pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.896756 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/875e9c09-a5e1-455a-b12e-e46ea43be9ed-utilities\") pod \"certified-operators-w7nfj\" (UID: \"875e9c09-a5e1-455a-b12e-e46ea43be9ed\") " pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.897288 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/875e9c09-a5e1-455a-b12e-e46ea43be9ed-catalog-content\") pod \"certified-operators-w7nfj\" (UID: \"875e9c09-a5e1-455a-b12e-e46ea43be9ed\") " pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.897389 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/875e9c09-a5e1-455a-b12e-e46ea43be9ed-utilities\") pod \"certified-operators-w7nfj\" (UID: \"875e9c09-a5e1-455a-b12e-e46ea43be9ed\") " pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.922109 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq8g9\" (UniqueName: \"kubernetes.io/projected/875e9c09-a5e1-455a-b12e-e46ea43be9ed-kube-api-access-tq8g9\") pod \"certified-operators-w7nfj\" (UID: \"875e9c09-a5e1-455a-b12e-e46ea43be9ed\") " pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:01:01 crc kubenswrapper[4739]: I1201 16:01:01.962890 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:01:02 crc kubenswrapper[4739]: I1201 16:01:02.474131 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w7nfj"] Dec 01 16:01:02 crc kubenswrapper[4739]: W1201 16:01:02.474604 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod875e9c09_a5e1_455a_b12e_e46ea43be9ed.slice/crio-dac444c5593671b13f79c430bb7495aee4ac6fd412a0cc72d697af59ff0343f4 WatchSource:0}: Error finding container dac444c5593671b13f79c430bb7495aee4ac6fd412a0cc72d697af59ff0343f4: Status 404 returned error can't find the container with id dac444c5593671b13f79c430bb7495aee4ac6fd412a0cc72d697af59ff0343f4 Dec 01 16:01:03 crc kubenswrapper[4739]: I1201 16:01:03.186811 4739 generic.go:334] "Generic (PLEG): container finished" podID="875e9c09-a5e1-455a-b12e-e46ea43be9ed" containerID="4518238ed1c7272568388d3699d0f938dca110865554d0eada543bbe739f859a" exitCode=0 Dec 01 16:01:03 crc kubenswrapper[4739]: I1201 16:01:03.187230 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7nfj" event={"ID":"875e9c09-a5e1-455a-b12e-e46ea43be9ed","Type":"ContainerDied","Data":"4518238ed1c7272568388d3699d0f938dca110865554d0eada543bbe739f859a"} Dec 01 16:01:03 crc kubenswrapper[4739]: I1201 16:01:03.187275 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7nfj" event={"ID":"875e9c09-a5e1-455a-b12e-e46ea43be9ed","Type":"ContainerStarted","Data":"dac444c5593671b13f79c430bb7495aee4ac6fd412a0cc72d697af59ff0343f4"} Dec 01 16:01:03 crc kubenswrapper[4739]: I1201 16:01:03.190004 4739 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:01:04 crc kubenswrapper[4739]: I1201 16:01:04.203499 4739 generic.go:334] "Generic (PLEG): container finished" podID="8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008" containerID="23ac4332746c5b9177f7f68b33b945ca3f78f459442d4ab4c5fa0894a74dae43" exitCode=0 Dec 01 16:01:04 crc kubenswrapper[4739]: I1201 16:01:04.203637 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410081-vqlxx" event={"ID":"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008","Type":"ContainerDied","Data":"23ac4332746c5b9177f7f68b33b945ca3f78f459442d4ab4c5fa0894a74dae43"} Dec 01 16:01:05 crc kubenswrapper[4739]: I1201 16:01:05.620110 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:05 crc kubenswrapper[4739]: I1201 16:01:05.775229 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-fernet-keys\") pod \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " Dec 01 16:01:05 crc kubenswrapper[4739]: I1201 16:01:05.776062 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stf4q\" (UniqueName: \"kubernetes.io/projected/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-kube-api-access-stf4q\") pod \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " Dec 01 16:01:05 crc kubenswrapper[4739]: I1201 16:01:05.776132 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-combined-ca-bundle\") pod \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " Dec 01 16:01:05 crc kubenswrapper[4739]: I1201 16:01:05.776608 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-config-data\") pod \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\" (UID: \"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008\") " Dec 01 16:01:05 crc kubenswrapper[4739]: I1201 16:01:05.782786 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008" (UID: "8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:01:05 crc kubenswrapper[4739]: I1201 16:01:05.791698 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-kube-api-access-stf4q" (OuterVolumeSpecName: "kube-api-access-stf4q") pod "8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008" (UID: "8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008"). InnerVolumeSpecName "kube-api-access-stf4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:01:05 crc kubenswrapper[4739]: I1201 16:01:05.811269 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008" (UID: "8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:01:05 crc kubenswrapper[4739]: I1201 16:01:05.833047 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-config-data" (OuterVolumeSpecName: "config-data") pod "8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008" (UID: "8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:01:05 crc kubenswrapper[4739]: I1201 16:01:05.879279 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:01:05 crc kubenswrapper[4739]: I1201 16:01:05.879330 4739 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 16:01:05 crc kubenswrapper[4739]: I1201 16:01:05.879351 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stf4q\" (UniqueName: \"kubernetes.io/projected/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-kube-api-access-stf4q\") on node \"crc\" DevicePath \"\"" Dec 01 16:01:05 crc kubenswrapper[4739]: I1201 16:01:05.879371 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:01:06 crc kubenswrapper[4739]: I1201 16:01:06.225370 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410081-vqlxx" event={"ID":"8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008","Type":"ContainerDied","Data":"b0cd99bf7d4fb7e93595c25fecbe68686c3e7530f121adc4e31e679ade4c33a5"} Dec 01 16:01:06 crc kubenswrapper[4739]: I1201 16:01:06.225454 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0cd99bf7d4fb7e93595c25fecbe68686c3e7530f121adc4e31e679ade4c33a5" Dec 01 16:01:06 crc kubenswrapper[4739]: I1201 16:01:06.225516 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410081-vqlxx" Dec 01 16:01:09 crc kubenswrapper[4739]: I1201 16:01:09.258794 4739 generic.go:334] "Generic (PLEG): container finished" podID="875e9c09-a5e1-455a-b12e-e46ea43be9ed" containerID="452fbb2768d590796077dca8baefba28c0ff7e9cf3ec5c21f54de086bbba6070" exitCode=0 Dec 01 16:01:09 crc kubenswrapper[4739]: I1201 16:01:09.258865 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7nfj" event={"ID":"875e9c09-a5e1-455a-b12e-e46ea43be9ed","Type":"ContainerDied","Data":"452fbb2768d590796077dca8baefba28c0ff7e9cf3ec5c21f54de086bbba6070"} Dec 01 16:01:13 crc kubenswrapper[4739]: I1201 16:01:13.477445 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:01:13 crc kubenswrapper[4739]: E1201 16:01:13.478512 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:01:16 crc kubenswrapper[4739]: I1201 16:01:16.327379 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7nfj" event={"ID":"875e9c09-a5e1-455a-b12e-e46ea43be9ed","Type":"ContainerStarted","Data":"9c2dcb5020e4c998e852d33aee7b8d35efeba821df91601ea0cf7d51d8c681f3"} Dec 01 16:01:16 crc kubenswrapper[4739]: I1201 16:01:16.349089 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w7nfj" podStartSLOduration=3.115413467 podStartE2EDuration="15.349069229s" podCreationTimestamp="2025-12-01 16:01:01 +0000 UTC" firstStartedPulling="2025-12-01 16:01:03.189474798 +0000 UTC m=+1565.015220932" lastFinishedPulling="2025-12-01 16:01:15.42313059 +0000 UTC m=+1577.248876694" observedRunningTime="2025-12-01 16:01:16.342778135 +0000 UTC m=+1578.168524229" watchObservedRunningTime="2025-12-01 16:01:16.349069229 +0000 UTC m=+1578.174815323" Dec 01 16:01:20 crc kubenswrapper[4739]: I1201 16:01:20.763371 4739 scope.go:117] "RemoveContainer" containerID="0b432b1274db4acbebc79ac56068c6a2a2fc8a58426387136bdb5c2b3a828ab2" Dec 01 16:01:21 crc kubenswrapper[4739]: I1201 16:01:21.963765 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:01:21 crc kubenswrapper[4739]: I1201 16:01:21.964150 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:01:22 crc kubenswrapper[4739]: I1201 16:01:22.030873 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:01:22 crc kubenswrapper[4739]: I1201 16:01:22.445981 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:01:22 crc kubenswrapper[4739]: I1201 16:01:22.542390 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w7nfj"] Dec 01 16:01:22 crc kubenswrapper[4739]: I1201 16:01:22.602548 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9jfkx"] Dec 01 16:01:22 crc kubenswrapper[4739]: I1201 16:01:22.602841 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9jfkx" podUID="c75bb144-38de-46c1-af46-26a0840c1632" containerName="registry-server" containerID="cri-o://5e04d551c5e71170b416dcb611a70084a5a0d915e93c76384f774a86d0f2250b" gracePeriod=2 Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.109006 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.224819 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c75bb144-38de-46c1-af46-26a0840c1632-catalog-content\") pod \"c75bb144-38de-46c1-af46-26a0840c1632\" (UID: \"c75bb144-38de-46c1-af46-26a0840c1632\") " Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.225061 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c75bb144-38de-46c1-af46-26a0840c1632-utilities\") pod \"c75bb144-38de-46c1-af46-26a0840c1632\" (UID: \"c75bb144-38de-46c1-af46-26a0840c1632\") " Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.225147 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqvnj\" (UniqueName: \"kubernetes.io/projected/c75bb144-38de-46c1-af46-26a0840c1632-kube-api-access-zqvnj\") pod \"c75bb144-38de-46c1-af46-26a0840c1632\" (UID: \"c75bb144-38de-46c1-af46-26a0840c1632\") " Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.225518 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c75bb144-38de-46c1-af46-26a0840c1632-utilities" (OuterVolumeSpecName: "utilities") pod "c75bb144-38de-46c1-af46-26a0840c1632" (UID: "c75bb144-38de-46c1-af46-26a0840c1632"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.226001 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c75bb144-38de-46c1-af46-26a0840c1632-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.231588 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c75bb144-38de-46c1-af46-26a0840c1632-kube-api-access-zqvnj" (OuterVolumeSpecName: "kube-api-access-zqvnj") pod "c75bb144-38de-46c1-af46-26a0840c1632" (UID: "c75bb144-38de-46c1-af46-26a0840c1632"). InnerVolumeSpecName "kube-api-access-zqvnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.268877 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c75bb144-38de-46c1-af46-26a0840c1632-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c75bb144-38de-46c1-af46-26a0840c1632" (UID: "c75bb144-38de-46c1-af46-26a0840c1632"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.327700 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqvnj\" (UniqueName: \"kubernetes.io/projected/c75bb144-38de-46c1-af46-26a0840c1632-kube-api-access-zqvnj\") on node \"crc\" DevicePath \"\"" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.327740 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c75bb144-38de-46c1-af46-26a0840c1632-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.395033 4739 generic.go:334] "Generic (PLEG): container finished" podID="c75bb144-38de-46c1-af46-26a0840c1632" containerID="5e04d551c5e71170b416dcb611a70084a5a0d915e93c76384f774a86d0f2250b" exitCode=0 Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.395097 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jfkx" event={"ID":"c75bb144-38de-46c1-af46-26a0840c1632","Type":"ContainerDied","Data":"5e04d551c5e71170b416dcb611a70084a5a0d915e93c76384f774a86d0f2250b"} Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.395146 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9jfkx" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.395170 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jfkx" event={"ID":"c75bb144-38de-46c1-af46-26a0840c1632","Type":"ContainerDied","Data":"f61198468e66eb5397719799d2e765b393abfb227f84514cee345be3210b0c07"} Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.395191 4739 scope.go:117] "RemoveContainer" containerID="5e04d551c5e71170b416dcb611a70084a5a0d915e93c76384f774a86d0f2250b" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.420711 4739 scope.go:117] "RemoveContainer" containerID="b9fd5198001b68bbdd90bf678024d7611af26b45788e6ad924f74c234412bfc4" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.433082 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9jfkx"] Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.453216 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9jfkx"] Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.470479 4739 scope.go:117] "RemoveContainer" containerID="69e528e161294d1152ab13b170a3e95cb478bf3d043d7ec43068059a3cbfddf4" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.500906 4739 scope.go:117] "RemoveContainer" containerID="5e04d551c5e71170b416dcb611a70084a5a0d915e93c76384f774a86d0f2250b" Dec 01 16:01:23 crc kubenswrapper[4739]: E1201 16:01:23.501409 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e04d551c5e71170b416dcb611a70084a5a0d915e93c76384f774a86d0f2250b\": container with ID starting with 5e04d551c5e71170b416dcb611a70084a5a0d915e93c76384f774a86d0f2250b not found: ID does not exist" containerID="5e04d551c5e71170b416dcb611a70084a5a0d915e93c76384f774a86d0f2250b" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.501473 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e04d551c5e71170b416dcb611a70084a5a0d915e93c76384f774a86d0f2250b"} err="failed to get container status \"5e04d551c5e71170b416dcb611a70084a5a0d915e93c76384f774a86d0f2250b\": rpc error: code = NotFound desc = could not find container \"5e04d551c5e71170b416dcb611a70084a5a0d915e93c76384f774a86d0f2250b\": container with ID starting with 5e04d551c5e71170b416dcb611a70084a5a0d915e93c76384f774a86d0f2250b not found: ID does not exist" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.501498 4739 scope.go:117] "RemoveContainer" containerID="b9fd5198001b68bbdd90bf678024d7611af26b45788e6ad924f74c234412bfc4" Dec 01 16:01:23 crc kubenswrapper[4739]: E1201 16:01:23.501912 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9fd5198001b68bbdd90bf678024d7611af26b45788e6ad924f74c234412bfc4\": container with ID starting with b9fd5198001b68bbdd90bf678024d7611af26b45788e6ad924f74c234412bfc4 not found: ID does not exist" containerID="b9fd5198001b68bbdd90bf678024d7611af26b45788e6ad924f74c234412bfc4" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.502044 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9fd5198001b68bbdd90bf678024d7611af26b45788e6ad924f74c234412bfc4"} err="failed to get container status \"b9fd5198001b68bbdd90bf678024d7611af26b45788e6ad924f74c234412bfc4\": rpc error: code = NotFound desc = could not find container \"b9fd5198001b68bbdd90bf678024d7611af26b45788e6ad924f74c234412bfc4\": container with ID starting with b9fd5198001b68bbdd90bf678024d7611af26b45788e6ad924f74c234412bfc4 not found: ID does not exist" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.502074 4739 scope.go:117] "RemoveContainer" containerID="69e528e161294d1152ab13b170a3e95cb478bf3d043d7ec43068059a3cbfddf4" Dec 01 16:01:23 crc kubenswrapper[4739]: E1201 16:01:23.502700 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69e528e161294d1152ab13b170a3e95cb478bf3d043d7ec43068059a3cbfddf4\": container with ID starting with 69e528e161294d1152ab13b170a3e95cb478bf3d043d7ec43068059a3cbfddf4 not found: ID does not exist" containerID="69e528e161294d1152ab13b170a3e95cb478bf3d043d7ec43068059a3cbfddf4" Dec 01 16:01:23 crc kubenswrapper[4739]: I1201 16:01:23.502768 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69e528e161294d1152ab13b170a3e95cb478bf3d043d7ec43068059a3cbfddf4"} err="failed to get container status \"69e528e161294d1152ab13b170a3e95cb478bf3d043d7ec43068059a3cbfddf4\": rpc error: code = NotFound desc = could not find container \"69e528e161294d1152ab13b170a3e95cb478bf3d043d7ec43068059a3cbfddf4\": container with ID starting with 69e528e161294d1152ab13b170a3e95cb478bf3d043d7ec43068059a3cbfddf4 not found: ID does not exist" Dec 01 16:01:24 crc kubenswrapper[4739]: I1201 16:01:24.490778 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c75bb144-38de-46c1-af46-26a0840c1632" path="/var/lib/kubelet/pods/c75bb144-38de-46c1-af46-26a0840c1632/volumes" Dec 01 16:01:28 crc kubenswrapper[4739]: I1201 16:01:28.504963 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:01:28 crc kubenswrapper[4739]: E1201 16:01:28.506064 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:01:42 crc kubenswrapper[4739]: I1201 16:01:42.478071 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:01:42 crc kubenswrapper[4739]: E1201 16:01:42.479501 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:01:44 crc kubenswrapper[4739]: I1201 16:01:44.640721 4739 generic.go:334] "Generic (PLEG): container finished" podID="5c7d3cbb-19d3-4855-8ad4-72504ba66787" containerID="3fac8a90bed2abfaa40fef7599e3db6a3a00c3706112ac25c5890105b83ed039" exitCode=0 Dec 01 16:01:44 crc kubenswrapper[4739]: I1201 16:01:44.640847 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" event={"ID":"5c7d3cbb-19d3-4855-8ad4-72504ba66787","Type":"ContainerDied","Data":"3fac8a90bed2abfaa40fef7599e3db6a3a00c3706112ac25c5890105b83ed039"} Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.138104 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.181810 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c7d3cbb-19d3-4855-8ad4-72504ba66787-ssh-key\") pod \"5c7d3cbb-19d3-4855-8ad4-72504ba66787\" (UID: \"5c7d3cbb-19d3-4855-8ad4-72504ba66787\") " Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.182106 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86msw\" (UniqueName: \"kubernetes.io/projected/5c7d3cbb-19d3-4855-8ad4-72504ba66787-kube-api-access-86msw\") pod \"5c7d3cbb-19d3-4855-8ad4-72504ba66787\" (UID: \"5c7d3cbb-19d3-4855-8ad4-72504ba66787\") " Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.182157 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c7d3cbb-19d3-4855-8ad4-72504ba66787-inventory\") pod \"5c7d3cbb-19d3-4855-8ad4-72504ba66787\" (UID: \"5c7d3cbb-19d3-4855-8ad4-72504ba66787\") " Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.189844 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c7d3cbb-19d3-4855-8ad4-72504ba66787-kube-api-access-86msw" (OuterVolumeSpecName: "kube-api-access-86msw") pod "5c7d3cbb-19d3-4855-8ad4-72504ba66787" (UID: "5c7d3cbb-19d3-4855-8ad4-72504ba66787"). InnerVolumeSpecName "kube-api-access-86msw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.212638 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c7d3cbb-19d3-4855-8ad4-72504ba66787-inventory" (OuterVolumeSpecName: "inventory") pod "5c7d3cbb-19d3-4855-8ad4-72504ba66787" (UID: "5c7d3cbb-19d3-4855-8ad4-72504ba66787"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.217554 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c7d3cbb-19d3-4855-8ad4-72504ba66787-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5c7d3cbb-19d3-4855-8ad4-72504ba66787" (UID: "5c7d3cbb-19d3-4855-8ad4-72504ba66787"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.284210 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86msw\" (UniqueName: \"kubernetes.io/projected/5c7d3cbb-19d3-4855-8ad4-72504ba66787-kube-api-access-86msw\") on node \"crc\" DevicePath \"\"" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.284250 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c7d3cbb-19d3-4855-8ad4-72504ba66787-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.284263 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c7d3cbb-19d3-4855-8ad4-72504ba66787-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.660254 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" event={"ID":"5c7d3cbb-19d3-4855-8ad4-72504ba66787","Type":"ContainerDied","Data":"a9c970c851d4a98da0388150708d0b25d357f5e7c7e769e7c67f720160e381e3"} Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.660611 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9c970c851d4a98da0388150708d0b25d357f5e7c7e769e7c67f720160e381e3" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.660311 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.751066 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx"] Dec 01 16:01:46 crc kubenswrapper[4739]: E1201 16:01:46.751554 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c75bb144-38de-46c1-af46-26a0840c1632" containerName="extract-content" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.751573 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="c75bb144-38de-46c1-af46-26a0840c1632" containerName="extract-content" Dec 01 16:01:46 crc kubenswrapper[4739]: E1201 16:01:46.751600 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c7d3cbb-19d3-4855-8ad4-72504ba66787" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.751609 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c7d3cbb-19d3-4855-8ad4-72504ba66787" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 16:01:46 crc kubenswrapper[4739]: E1201 16:01:46.751626 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c75bb144-38de-46c1-af46-26a0840c1632" containerName="extract-utilities" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.751634 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="c75bb144-38de-46c1-af46-26a0840c1632" containerName="extract-utilities" Dec 01 16:01:46 crc kubenswrapper[4739]: E1201 16:01:46.751650 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008" containerName="keystone-cron" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.751658 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008" containerName="keystone-cron" Dec 01 16:01:46 crc kubenswrapper[4739]: E1201 16:01:46.751687 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c75bb144-38de-46c1-af46-26a0840c1632" containerName="registry-server" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.751695 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="c75bb144-38de-46c1-af46-26a0840c1632" containerName="registry-server" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.751914 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c7d3cbb-19d3-4855-8ad4-72504ba66787" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.751939 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="c75bb144-38de-46c1-af46-26a0840c1632" containerName="registry-server" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.751953 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008" containerName="keystone-cron" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.752670 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.754941 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.755318 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.755690 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.756011 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.765168 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx"] Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.792388 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b8542fe-a173-43b5-9654-89e34562f4f5-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx\" (UID: \"0b8542fe-a173-43b5-9654-89e34562f4f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.792577 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b8542fe-a173-43b5-9654-89e34562f4f5-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx\" (UID: \"0b8542fe-a173-43b5-9654-89e34562f4f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.792623 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9gf4\" (UniqueName: \"kubernetes.io/projected/0b8542fe-a173-43b5-9654-89e34562f4f5-kube-api-access-h9gf4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx\" (UID: \"0b8542fe-a173-43b5-9654-89e34562f4f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.894083 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b8542fe-a173-43b5-9654-89e34562f4f5-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx\" (UID: \"0b8542fe-a173-43b5-9654-89e34562f4f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.894192 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b8542fe-a173-43b5-9654-89e34562f4f5-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx\" (UID: \"0b8542fe-a173-43b5-9654-89e34562f4f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.894239 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9gf4\" (UniqueName: \"kubernetes.io/projected/0b8542fe-a173-43b5-9654-89e34562f4f5-kube-api-access-h9gf4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx\" (UID: \"0b8542fe-a173-43b5-9654-89e34562f4f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.898372 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b8542fe-a173-43b5-9654-89e34562f4f5-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx\" (UID: \"0b8542fe-a173-43b5-9654-89e34562f4f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.898973 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b8542fe-a173-43b5-9654-89e34562f4f5-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx\" (UID: \"0b8542fe-a173-43b5-9654-89e34562f4f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" Dec 01 16:01:46 crc kubenswrapper[4739]: I1201 16:01:46.912977 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9gf4\" (UniqueName: \"kubernetes.io/projected/0b8542fe-a173-43b5-9654-89e34562f4f5-kube-api-access-h9gf4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx\" (UID: \"0b8542fe-a173-43b5-9654-89e34562f4f5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" Dec 01 16:01:47 crc kubenswrapper[4739]: I1201 16:01:47.084264 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" Dec 01 16:01:47 crc kubenswrapper[4739]: I1201 16:01:47.650430 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx"] Dec 01 16:01:47 crc kubenswrapper[4739]: I1201 16:01:47.669688 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" event={"ID":"0b8542fe-a173-43b5-9654-89e34562f4f5","Type":"ContainerStarted","Data":"f20e77e94ba7f9d2f09b7304cc6c9e81f6a18b65e10ffb2a141b024dd690fa4d"} Dec 01 16:01:49 crc kubenswrapper[4739]: I1201 16:01:49.700458 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" event={"ID":"0b8542fe-a173-43b5-9654-89e34562f4f5","Type":"ContainerStarted","Data":"1ce022649232a707c748a1a9afd1480938b067c019eec773ab1b09bdc08f8a1c"} Dec 01 16:01:54 crc kubenswrapper[4739]: I1201 16:01:54.756316 4739 generic.go:334] "Generic (PLEG): container finished" podID="0b8542fe-a173-43b5-9654-89e34562f4f5" containerID="1ce022649232a707c748a1a9afd1480938b067c019eec773ab1b09bdc08f8a1c" exitCode=0 Dec 01 16:01:54 crc kubenswrapper[4739]: I1201 16:01:54.756395 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" event={"ID":"0b8542fe-a173-43b5-9654-89e34562f4f5","Type":"ContainerDied","Data":"1ce022649232a707c748a1a9afd1480938b067c019eec773ab1b09bdc08f8a1c"} Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.269997 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.298303 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b8542fe-a173-43b5-9654-89e34562f4f5-ssh-key\") pod \"0b8542fe-a173-43b5-9654-89e34562f4f5\" (UID: \"0b8542fe-a173-43b5-9654-89e34562f4f5\") " Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.300089 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b8542fe-a173-43b5-9654-89e34562f4f5-inventory\") pod \"0b8542fe-a173-43b5-9654-89e34562f4f5\" (UID: \"0b8542fe-a173-43b5-9654-89e34562f4f5\") " Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.300143 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9gf4\" (UniqueName: \"kubernetes.io/projected/0b8542fe-a173-43b5-9654-89e34562f4f5-kube-api-access-h9gf4\") pod \"0b8542fe-a173-43b5-9654-89e34562f4f5\" (UID: \"0b8542fe-a173-43b5-9654-89e34562f4f5\") " Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.323429 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b8542fe-a173-43b5-9654-89e34562f4f5-kube-api-access-h9gf4" (OuterVolumeSpecName: "kube-api-access-h9gf4") pod "0b8542fe-a173-43b5-9654-89e34562f4f5" (UID: "0b8542fe-a173-43b5-9654-89e34562f4f5"). InnerVolumeSpecName "kube-api-access-h9gf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.360007 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b8542fe-a173-43b5-9654-89e34562f4f5-inventory" (OuterVolumeSpecName: "inventory") pod "0b8542fe-a173-43b5-9654-89e34562f4f5" (UID: "0b8542fe-a173-43b5-9654-89e34562f4f5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.360144 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b8542fe-a173-43b5-9654-89e34562f4f5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0b8542fe-a173-43b5-9654-89e34562f4f5" (UID: "0b8542fe-a173-43b5-9654-89e34562f4f5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.405289 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b8542fe-a173-43b5-9654-89e34562f4f5-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.405331 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9gf4\" (UniqueName: \"kubernetes.io/projected/0b8542fe-a173-43b5-9654-89e34562f4f5-kube-api-access-h9gf4\") on node \"crc\" DevicePath \"\"" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.405342 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b8542fe-a173-43b5-9654-89e34562f4f5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.776547 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" event={"ID":"0b8542fe-a173-43b5-9654-89e34562f4f5","Type":"ContainerDied","Data":"f20e77e94ba7f9d2f09b7304cc6c9e81f6a18b65e10ffb2a141b024dd690fa4d"} Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.776596 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f20e77e94ba7f9d2f09b7304cc6c9e81f6a18b65e10ffb2a141b024dd690fa4d" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.776635 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.875782 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9"] Dec 01 16:01:56 crc kubenswrapper[4739]: E1201 16:01:56.876204 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b8542fe-a173-43b5-9654-89e34562f4f5" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.876222 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b8542fe-a173-43b5-9654-89e34562f4f5" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.876409 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b8542fe-a173-43b5-9654-89e34562f4f5" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.877034 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.881081 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.882000 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.883640 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.883715 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.891396 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9"] Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.915578 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz7k5\" (UniqueName: \"kubernetes.io/projected/b5a13eff-ceca-4a92-8cf0-8334c07e3828-kube-api-access-rz7k5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-l4rv9\" (UID: \"b5a13eff-ceca-4a92-8cf0-8334c07e3828\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.915643 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5a13eff-ceca-4a92-8cf0-8334c07e3828-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-l4rv9\" (UID: \"b5a13eff-ceca-4a92-8cf0-8334c07e3828\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" Dec 01 16:01:56 crc kubenswrapper[4739]: I1201 16:01:56.915827 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5a13eff-ceca-4a92-8cf0-8334c07e3828-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-l4rv9\" (UID: \"b5a13eff-ceca-4a92-8cf0-8334c07e3828\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.017639 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz7k5\" (UniqueName: \"kubernetes.io/projected/b5a13eff-ceca-4a92-8cf0-8334c07e3828-kube-api-access-rz7k5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-l4rv9\" (UID: \"b5a13eff-ceca-4a92-8cf0-8334c07e3828\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.017713 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5a13eff-ceca-4a92-8cf0-8334c07e3828-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-l4rv9\" (UID: \"b5a13eff-ceca-4a92-8cf0-8334c07e3828\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.017810 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5a13eff-ceca-4a92-8cf0-8334c07e3828-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-l4rv9\" (UID: \"b5a13eff-ceca-4a92-8cf0-8334c07e3828\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.023403 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5a13eff-ceca-4a92-8cf0-8334c07e3828-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-l4rv9\" (UID: \"b5a13eff-ceca-4a92-8cf0-8334c07e3828\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.026020 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5a13eff-ceca-4a92-8cf0-8334c07e3828-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-l4rv9\" (UID: \"b5a13eff-ceca-4a92-8cf0-8334c07e3828\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.042456 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz7k5\" (UniqueName: \"kubernetes.io/projected/b5a13eff-ceca-4a92-8cf0-8334c07e3828-kube-api-access-rz7k5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-l4rv9\" (UID: \"b5a13eff-ceca-4a92-8cf0-8334c07e3828\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.052734 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-dxcwt"] Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.068312 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-8170-account-create-update-8jz42"] Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.079323 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-ed22-account-create-update-c4lxx"] Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.087157 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-d1ed-account-create-update-64cxx"] Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.094760 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-dxcwt"] Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.102547 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-skpk2"] Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.111585 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-8170-account-create-update-8jz42"] Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.120463 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-skpk2"] Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.130766 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-ed22-account-create-update-c4lxx"] Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.140600 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-d1ed-account-create-update-64cxx"] Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.197846 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.478026 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:01:57 crc kubenswrapper[4739]: E1201 16:01:57.478307 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.712026 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9"] Dec 01 16:01:57 crc kubenswrapper[4739]: I1201 16:01:57.785941 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" event={"ID":"b5a13eff-ceca-4a92-8cf0-8334c07e3828","Type":"ContainerStarted","Data":"d12368de77c3b00d84219b39076511fa260bf0dccf28739d4c745942f34e4def"} Dec 01 16:01:58 crc kubenswrapper[4739]: I1201 16:01:58.028565 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-27pv4"] Dec 01 16:01:58 crc kubenswrapper[4739]: I1201 16:01:58.039526 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-27pv4"] Dec 01 16:01:58 crc kubenswrapper[4739]: I1201 16:01:58.491196 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52f45a77-0faf-42ae-b052-ccb63b092440" path="/var/lib/kubelet/pods/52f45a77-0faf-42ae-b052-ccb63b092440/volumes" Dec 01 16:01:58 crc kubenswrapper[4739]: I1201 16:01:58.492395 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="618420cf-adef-4233-b155-d213621354a0" path="/var/lib/kubelet/pods/618420cf-adef-4233-b155-d213621354a0/volumes" Dec 01 16:01:58 crc kubenswrapper[4739]: I1201 16:01:58.493038 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9add3351-b3aa-41a1-8c08-5c10147e2520" path="/var/lib/kubelet/pods/9add3351-b3aa-41a1-8c08-5c10147e2520/volumes" Dec 01 16:01:58 crc kubenswrapper[4739]: I1201 16:01:58.493688 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5e17350-562f-4cd1-a861-0cea798c8528" path="/var/lib/kubelet/pods/d5e17350-562f-4cd1-a861-0cea798c8528/volumes" Dec 01 16:01:58 crc kubenswrapper[4739]: I1201 16:01:58.494799 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dceccaf5-f087-477b-b9ac-4a0748136112" path="/var/lib/kubelet/pods/dceccaf5-f087-477b-b9ac-4a0748136112/volumes" Dec 01 16:01:58 crc kubenswrapper[4739]: I1201 16:01:58.495373 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7ff093b-772b-4c09-ba41-6f5fd9324591" path="/var/lib/kubelet/pods/f7ff093b-772b-4c09-ba41-6f5fd9324591/volumes" Dec 01 16:01:58 crc kubenswrapper[4739]: I1201 16:01:58.796126 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" event={"ID":"b5a13eff-ceca-4a92-8cf0-8334c07e3828","Type":"ContainerStarted","Data":"f804a154b9c41c7489a2b8af6e2aa383bf99e78ecd75cb2c7f48ee8c24c89faf"} Dec 01 16:01:58 crc kubenswrapper[4739]: I1201 16:01:58.818787 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" podStartSLOduration=2.296917882 podStartE2EDuration="2.818768456s" podCreationTimestamp="2025-12-01 16:01:56 +0000 UTC" firstStartedPulling="2025-12-01 16:01:57.712532723 +0000 UTC m=+1619.538278827" lastFinishedPulling="2025-12-01 16:01:58.234383267 +0000 UTC m=+1620.060129401" observedRunningTime="2025-12-01 16:01:58.811652006 +0000 UTC m=+1620.637398120" watchObservedRunningTime="2025-12-01 16:01:58.818768456 +0000 UTC m=+1620.644514550" Dec 01 16:02:08 crc kubenswrapper[4739]: I1201 16:02:08.488392 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:02:08 crc kubenswrapper[4739]: E1201 16:02:08.490752 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:02:20 crc kubenswrapper[4739]: I1201 16:02:20.833786 4739 scope.go:117] "RemoveContainer" containerID="b18a60b2ae4f8812e5b38987deb886bf2bc8e2925c5afbc43170c982e5f56c09" Dec 01 16:02:20 crc kubenswrapper[4739]: I1201 16:02:20.875096 4739 scope.go:117] "RemoveContainer" containerID="d2f6a3cb0b76bdca0eab5cbaa0f63fe887ee181aa1f33abab0300196c6bc4c40" Dec 01 16:02:20 crc kubenswrapper[4739]: I1201 16:02:20.950920 4739 scope.go:117] "RemoveContainer" containerID="1f66a2f72a31f5b7dc08bf035004c414c9c2d6d960dde14430a819e6c6ec1c38" Dec 01 16:02:21 crc kubenswrapper[4739]: I1201 16:02:21.018514 4739 scope.go:117] "RemoveContainer" containerID="d22e2824bf982944c902071f019940b381768974cc86e288e7fb4033793dc037" Dec 01 16:02:21 crc kubenswrapper[4739]: I1201 16:02:21.055918 4739 scope.go:117] "RemoveContainer" containerID="9eb8e3ce1a5302ca942a68401089eaede902fc7a45cee116ab78056a8aad1124" Dec 01 16:02:21 crc kubenswrapper[4739]: I1201 16:02:21.109169 4739 scope.go:117] "RemoveContainer" containerID="52a711a736dffeb17d2be406d7ef5ba5f2306a1e9a8afd8f83bc9db87ce8c990" Dec 01 16:02:22 crc kubenswrapper[4739]: I1201 16:02:22.065502 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-8zdxf"] Dec 01 16:02:22 crc kubenswrapper[4739]: I1201 16:02:22.100056 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-8zdxf"] Dec 01 16:02:22 crc kubenswrapper[4739]: I1201 16:02:22.478781 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:02:22 crc kubenswrapper[4739]: E1201 16:02:22.479188 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:02:22 crc kubenswrapper[4739]: I1201 16:02:22.495940 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc6eb674-5455-4596-8a2a-5b562dd87ef3" path="/var/lib/kubelet/pods/bc6eb674-5455-4596-8a2a-5b562dd87ef3/volumes" Dec 01 16:02:34 crc kubenswrapper[4739]: I1201 16:02:34.051240 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-6sn86"] Dec 01 16:02:34 crc kubenswrapper[4739]: I1201 16:02:34.065884 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-1814-account-create-update-7l5hz"] Dec 01 16:02:34 crc kubenswrapper[4739]: I1201 16:02:34.082448 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-4jbjk"] Dec 01 16:02:34 crc kubenswrapper[4739]: I1201 16:02:34.093778 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-1814-account-create-update-7l5hz"] Dec 01 16:02:34 crc kubenswrapper[4739]: I1201 16:02:34.103026 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-6sn86"] Dec 01 16:02:34 crc kubenswrapper[4739]: I1201 16:02:34.112025 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-4jbjk"] Dec 01 16:02:34 crc kubenswrapper[4739]: I1201 16:02:34.486039 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56bbe10e-0b3e-48c5-a806-cf5f1dec0935" path="/var/lib/kubelet/pods/56bbe10e-0b3e-48c5-a806-cf5f1dec0935/volumes" Dec 01 16:02:34 crc kubenswrapper[4739]: I1201 16:02:34.486591 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58e52c5d-7e41-4145-abad-945e9cb9d01b" path="/var/lib/kubelet/pods/58e52c5d-7e41-4145-abad-945e9cb9d01b/volumes" Dec 01 16:02:34 crc kubenswrapper[4739]: I1201 16:02:34.487124 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dec9942-68fd-4595-826c-9cc4bace11aa" path="/var/lib/kubelet/pods/6dec9942-68fd-4595-826c-9cc4bace11aa/volumes" Dec 01 16:02:35 crc kubenswrapper[4739]: I1201 16:02:35.029831 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-9373-account-create-update-64mfz"] Dec 01 16:02:35 crc kubenswrapper[4739]: I1201 16:02:35.046151 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-f640-account-create-update-vwhg8"] Dec 01 16:02:35 crc kubenswrapper[4739]: I1201 16:02:35.055531 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-pv8tn"] Dec 01 16:02:35 crc kubenswrapper[4739]: I1201 16:02:35.062542 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-9373-account-create-update-64mfz"] Dec 01 16:02:35 crc kubenswrapper[4739]: I1201 16:02:35.069751 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-pv8tn"] Dec 01 16:02:35 crc kubenswrapper[4739]: I1201 16:02:35.076849 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-f640-account-create-update-vwhg8"] Dec 01 16:02:36 crc kubenswrapper[4739]: I1201 16:02:36.496732 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bec2f8e-e989-4408-918b-3b9ad6f01af8" path="/var/lib/kubelet/pods/1bec2f8e-e989-4408-918b-3b9ad6f01af8/volumes" Dec 01 16:02:36 crc kubenswrapper[4739]: I1201 16:02:36.498742 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b93f1ae-e632-4bca-a89e-3ce44858a60e" path="/var/lib/kubelet/pods/5b93f1ae-e632-4bca-a89e-3ce44858a60e/volumes" Dec 01 16:02:36 crc kubenswrapper[4739]: I1201 16:02:36.500709 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a16816d-ed5f-4324-97d5-c30fb6b381bd" path="/var/lib/kubelet/pods/9a16816d-ed5f-4324-97d5-c30fb6b381bd/volumes" Dec 01 16:02:37 crc kubenswrapper[4739]: I1201 16:02:37.477213 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:02:37 crc kubenswrapper[4739]: E1201 16:02:37.477758 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:02:41 crc kubenswrapper[4739]: I1201 16:02:41.217669 4739 generic.go:334] "Generic (PLEG): container finished" podID="b5a13eff-ceca-4a92-8cf0-8334c07e3828" containerID="f804a154b9c41c7489a2b8af6e2aa383bf99e78ecd75cb2c7f48ee8c24c89faf" exitCode=0 Dec 01 16:02:41 crc kubenswrapper[4739]: I1201 16:02:41.217745 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" event={"ID":"b5a13eff-ceca-4a92-8cf0-8334c07e3828","Type":"ContainerDied","Data":"f804a154b9c41c7489a2b8af6e2aa383bf99e78ecd75cb2c7f48ee8c24c89faf"} Dec 01 16:02:42 crc kubenswrapper[4739]: I1201 16:02:42.787793 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" Dec 01 16:02:42 crc kubenswrapper[4739]: I1201 16:02:42.975633 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5a13eff-ceca-4a92-8cf0-8334c07e3828-inventory\") pod \"b5a13eff-ceca-4a92-8cf0-8334c07e3828\" (UID: \"b5a13eff-ceca-4a92-8cf0-8334c07e3828\") " Dec 01 16:02:42 crc kubenswrapper[4739]: I1201 16:02:42.975879 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5a13eff-ceca-4a92-8cf0-8334c07e3828-ssh-key\") pod \"b5a13eff-ceca-4a92-8cf0-8334c07e3828\" (UID: \"b5a13eff-ceca-4a92-8cf0-8334c07e3828\") " Dec 01 16:02:42 crc kubenswrapper[4739]: I1201 16:02:42.975998 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rz7k5\" (UniqueName: \"kubernetes.io/projected/b5a13eff-ceca-4a92-8cf0-8334c07e3828-kube-api-access-rz7k5\") pod \"b5a13eff-ceca-4a92-8cf0-8334c07e3828\" (UID: \"b5a13eff-ceca-4a92-8cf0-8334c07e3828\") " Dec 01 16:02:42 crc kubenswrapper[4739]: I1201 16:02:42.983711 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5a13eff-ceca-4a92-8cf0-8334c07e3828-kube-api-access-rz7k5" (OuterVolumeSpecName: "kube-api-access-rz7k5") pod "b5a13eff-ceca-4a92-8cf0-8334c07e3828" (UID: "b5a13eff-ceca-4a92-8cf0-8334c07e3828"). InnerVolumeSpecName "kube-api-access-rz7k5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.006653 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5a13eff-ceca-4a92-8cf0-8334c07e3828-inventory" (OuterVolumeSpecName: "inventory") pod "b5a13eff-ceca-4a92-8cf0-8334c07e3828" (UID: "b5a13eff-ceca-4a92-8cf0-8334c07e3828"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.009339 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5a13eff-ceca-4a92-8cf0-8334c07e3828-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b5a13eff-ceca-4a92-8cf0-8334c07e3828" (UID: "b5a13eff-ceca-4a92-8cf0-8334c07e3828"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.077936 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rz7k5\" (UniqueName: \"kubernetes.io/projected/b5a13eff-ceca-4a92-8cf0-8334c07e3828-kube-api-access-rz7k5\") on node \"crc\" DevicePath \"\"" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.077973 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5a13eff-ceca-4a92-8cf0-8334c07e3828-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.077988 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5a13eff-ceca-4a92-8cf0-8334c07e3828-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.243150 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" event={"ID":"b5a13eff-ceca-4a92-8cf0-8334c07e3828","Type":"ContainerDied","Data":"d12368de77c3b00d84219b39076511fa260bf0dccf28739d4c745942f34e4def"} Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.243377 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d12368de77c3b00d84219b39076511fa260bf0dccf28739d4c745942f34e4def" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.243247 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.360004 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn"] Dec 01 16:02:43 crc kubenswrapper[4739]: E1201 16:02:43.360729 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5a13eff-ceca-4a92-8cf0-8334c07e3828" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.360788 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5a13eff-ceca-4a92-8cf0-8334c07e3828" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.361204 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5a13eff-ceca-4a92-8cf0-8334c07e3828" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.362286 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.366133 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.366151 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.368380 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.368642 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.379785 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn"] Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.487232 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v25qq\" (UniqueName: \"kubernetes.io/projected/3027bd9d-3e05-44df-968e-228a275dcc54-kube-api-access-v25qq\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn\" (UID: \"3027bd9d-3e05-44df-968e-228a275dcc54\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.487309 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3027bd9d-3e05-44df-968e-228a275dcc54-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn\" (UID: \"3027bd9d-3e05-44df-968e-228a275dcc54\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.488117 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3027bd9d-3e05-44df-968e-228a275dcc54-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn\" (UID: \"3027bd9d-3e05-44df-968e-228a275dcc54\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.589945 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v25qq\" (UniqueName: \"kubernetes.io/projected/3027bd9d-3e05-44df-968e-228a275dcc54-kube-api-access-v25qq\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn\" (UID: \"3027bd9d-3e05-44df-968e-228a275dcc54\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.590235 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3027bd9d-3e05-44df-968e-228a275dcc54-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn\" (UID: \"3027bd9d-3e05-44df-968e-228a275dcc54\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.590373 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3027bd9d-3e05-44df-968e-228a275dcc54-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn\" (UID: \"3027bd9d-3e05-44df-968e-228a275dcc54\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.595519 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3027bd9d-3e05-44df-968e-228a275dcc54-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn\" (UID: \"3027bd9d-3e05-44df-968e-228a275dcc54\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.600220 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3027bd9d-3e05-44df-968e-228a275dcc54-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn\" (UID: \"3027bd9d-3e05-44df-968e-228a275dcc54\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.624908 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v25qq\" (UniqueName: \"kubernetes.io/projected/3027bd9d-3e05-44df-968e-228a275dcc54-kube-api-access-v25qq\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn\" (UID: \"3027bd9d-3e05-44df-968e-228a275dcc54\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" Dec 01 16:02:43 crc kubenswrapper[4739]: I1201 16:02:43.682744 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" Dec 01 16:02:44 crc kubenswrapper[4739]: I1201 16:02:44.276332 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn"] Dec 01 16:02:45 crc kubenswrapper[4739]: I1201 16:02:45.271848 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" event={"ID":"3027bd9d-3e05-44df-968e-228a275dcc54","Type":"ContainerStarted","Data":"577c84f6a4649de3d4b7a9eeca3414fbf7ae3bb3c7fe99e1e794d138ad80aab6"} Dec 01 16:02:45 crc kubenswrapper[4739]: I1201 16:02:45.272395 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" event={"ID":"3027bd9d-3e05-44df-968e-228a275dcc54","Type":"ContainerStarted","Data":"c02084efe4b4ef2160f00b285a7b8fbdd60240aeb1b45f7d3477d14658663bb8"} Dec 01 16:02:45 crc kubenswrapper[4739]: I1201 16:02:45.307203 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" podStartSLOduration=1.603754346 podStartE2EDuration="2.307174224s" podCreationTimestamp="2025-12-01 16:02:43 +0000 UTC" firstStartedPulling="2025-12-01 16:02:44.274474145 +0000 UTC m=+1666.100220249" lastFinishedPulling="2025-12-01 16:02:44.977893993 +0000 UTC m=+1666.803640127" observedRunningTime="2025-12-01 16:02:45.299404534 +0000 UTC m=+1667.125150668" watchObservedRunningTime="2025-12-01 16:02:45.307174224 +0000 UTC m=+1667.132920328" Dec 01 16:02:46 crc kubenswrapper[4739]: I1201 16:02:46.038114 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-fm2qz"] Dec 01 16:02:46 crc kubenswrapper[4739]: I1201 16:02:46.050004 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-fm2qz"] Dec 01 16:02:46 crc kubenswrapper[4739]: I1201 16:02:46.493952 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1" path="/var/lib/kubelet/pods/3bd00bf1-2bbd-41a7-9097-4b4237a5b6b1/volumes" Dec 01 16:02:49 crc kubenswrapper[4739]: I1201 16:02:49.478094 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:02:49 crc kubenswrapper[4739]: E1201 16:02:49.478911 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:02:50 crc kubenswrapper[4739]: I1201 16:02:50.331643 4739 generic.go:334] "Generic (PLEG): container finished" podID="3027bd9d-3e05-44df-968e-228a275dcc54" containerID="577c84f6a4649de3d4b7a9eeca3414fbf7ae3bb3c7fe99e1e794d138ad80aab6" exitCode=0 Dec 01 16:02:50 crc kubenswrapper[4739]: I1201 16:02:50.331759 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" event={"ID":"3027bd9d-3e05-44df-968e-228a275dcc54","Type":"ContainerDied","Data":"577c84f6a4649de3d4b7a9eeca3414fbf7ae3bb3c7fe99e1e794d138ad80aab6"} Dec 01 16:02:51 crc kubenswrapper[4739]: I1201 16:02:51.785642 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" Dec 01 16:02:51 crc kubenswrapper[4739]: I1201 16:02:51.881193 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3027bd9d-3e05-44df-968e-228a275dcc54-ssh-key\") pod \"3027bd9d-3e05-44df-968e-228a275dcc54\" (UID: \"3027bd9d-3e05-44df-968e-228a275dcc54\") " Dec 01 16:02:51 crc kubenswrapper[4739]: I1201 16:02:51.881339 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v25qq\" (UniqueName: \"kubernetes.io/projected/3027bd9d-3e05-44df-968e-228a275dcc54-kube-api-access-v25qq\") pod \"3027bd9d-3e05-44df-968e-228a275dcc54\" (UID: \"3027bd9d-3e05-44df-968e-228a275dcc54\") " Dec 01 16:02:51 crc kubenswrapper[4739]: I1201 16:02:51.881716 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3027bd9d-3e05-44df-968e-228a275dcc54-inventory\") pod \"3027bd9d-3e05-44df-968e-228a275dcc54\" (UID: \"3027bd9d-3e05-44df-968e-228a275dcc54\") " Dec 01 16:02:51 crc kubenswrapper[4739]: I1201 16:02:51.887459 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3027bd9d-3e05-44df-968e-228a275dcc54-kube-api-access-v25qq" (OuterVolumeSpecName: "kube-api-access-v25qq") pod "3027bd9d-3e05-44df-968e-228a275dcc54" (UID: "3027bd9d-3e05-44df-968e-228a275dcc54"). InnerVolumeSpecName "kube-api-access-v25qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:02:51 crc kubenswrapper[4739]: I1201 16:02:51.913460 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3027bd9d-3e05-44df-968e-228a275dcc54-inventory" (OuterVolumeSpecName: "inventory") pod "3027bd9d-3e05-44df-968e-228a275dcc54" (UID: "3027bd9d-3e05-44df-968e-228a275dcc54"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:02:51 crc kubenswrapper[4739]: I1201 16:02:51.913505 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3027bd9d-3e05-44df-968e-228a275dcc54-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3027bd9d-3e05-44df-968e-228a275dcc54" (UID: "3027bd9d-3e05-44df-968e-228a275dcc54"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:02:51 crc kubenswrapper[4739]: I1201 16:02:51.985075 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3027bd9d-3e05-44df-968e-228a275dcc54-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:02:51 crc kubenswrapper[4739]: I1201 16:02:51.985134 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3027bd9d-3e05-44df-968e-228a275dcc54-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:02:51 crc kubenswrapper[4739]: I1201 16:02:51.985155 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v25qq\" (UniqueName: \"kubernetes.io/projected/3027bd9d-3e05-44df-968e-228a275dcc54-kube-api-access-v25qq\") on node \"crc\" DevicePath \"\"" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.365741 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" event={"ID":"3027bd9d-3e05-44df-968e-228a275dcc54","Type":"ContainerDied","Data":"c02084efe4b4ef2160f00b285a7b8fbdd60240aeb1b45f7d3477d14658663bb8"} Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.366114 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c02084efe4b4ef2160f00b285a7b8fbdd60240aeb1b45f7d3477d14658663bb8" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.365818 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.455441 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd"] Dec 01 16:02:52 crc kubenswrapper[4739]: E1201 16:02:52.455953 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3027bd9d-3e05-44df-968e-228a275dcc54" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.455980 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="3027bd9d-3e05-44df-968e-228a275dcc54" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.456237 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="3027bd9d-3e05-44df-968e-228a275dcc54" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.457008 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.461753 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.461828 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.462137 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.462596 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.469305 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd"] Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.495874 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8qvk\" (UniqueName: \"kubernetes.io/projected/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-kube-api-access-f8qvk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd\" (UID: \"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.496095 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd\" (UID: \"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.496207 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd\" (UID: \"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.597604 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8qvk\" (UniqueName: \"kubernetes.io/projected/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-kube-api-access-f8qvk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd\" (UID: \"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.597672 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd\" (UID: \"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.597705 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd\" (UID: \"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.603779 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd\" (UID: \"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.609342 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd\" (UID: \"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.621533 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8qvk\" (UniqueName: \"kubernetes.io/projected/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-kube-api-access-f8qvk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd\" (UID: \"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" Dec 01 16:02:52 crc kubenswrapper[4739]: I1201 16:02:52.789954 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" Dec 01 16:02:53 crc kubenswrapper[4739]: I1201 16:02:53.119232 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd"] Dec 01 16:02:53 crc kubenswrapper[4739]: I1201 16:02:53.378936 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" event={"ID":"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9","Type":"ContainerStarted","Data":"81fd51f9bcc1443e8844b01ca99bfa3c64c39f60e67c5727ab877f5f09a48ce3"} Dec 01 16:02:55 crc kubenswrapper[4739]: I1201 16:02:55.413142 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" event={"ID":"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9","Type":"ContainerStarted","Data":"2d1187e9457eac46edf1a1c390e3a91f1575ec4d78e403ede67f59bc9d257eb6"} Dec 01 16:02:55 crc kubenswrapper[4739]: I1201 16:02:55.463074 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" podStartSLOduration=2.211667664 podStartE2EDuration="3.463044994s" podCreationTimestamp="2025-12-01 16:02:52 +0000 UTC" firstStartedPulling="2025-12-01 16:02:53.125914285 +0000 UTC m=+1674.951660389" lastFinishedPulling="2025-12-01 16:02:54.377291595 +0000 UTC m=+1676.203037719" observedRunningTime="2025-12-01 16:02:55.439050333 +0000 UTC m=+1677.264796467" watchObservedRunningTime="2025-12-01 16:02:55.463044994 +0000 UTC m=+1677.288791128" Dec 01 16:03:01 crc kubenswrapper[4739]: I1201 16:03:01.476821 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:03:01 crc kubenswrapper[4739]: E1201 16:03:01.477777 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:03:09 crc kubenswrapper[4739]: I1201 16:03:09.082003 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-28jb8"] Dec 01 16:03:09 crc kubenswrapper[4739]: I1201 16:03:09.095828 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-28jb8"] Dec 01 16:03:10 crc kubenswrapper[4739]: I1201 16:03:10.494663 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2999b95-023b-479c-a18c-0646e3e414d2" path="/var/lib/kubelet/pods/c2999b95-023b-479c-a18c-0646e3e414d2/volumes" Dec 01 16:03:13 crc kubenswrapper[4739]: I1201 16:03:13.045309 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-2lnl5"] Dec 01 16:03:13 crc kubenswrapper[4739]: I1201 16:03:13.057281 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-2lnl5"] Dec 01 16:03:14 crc kubenswrapper[4739]: I1201 16:03:14.477378 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:03:14 crc kubenswrapper[4739]: E1201 16:03:14.478300 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:03:14 crc kubenswrapper[4739]: I1201 16:03:14.498168 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baaebe35-daa8-449a-85fc-069056a1e9bc" path="/var/lib/kubelet/pods/baaebe35-daa8-449a-85fc-069056a1e9bc/volumes" Dec 01 16:03:20 crc kubenswrapper[4739]: I1201 16:03:20.050111 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-ztf2h"] Dec 01 16:03:20 crc kubenswrapper[4739]: I1201 16:03:20.065447 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-ztf2h"] Dec 01 16:03:20 crc kubenswrapper[4739]: I1201 16:03:20.495130 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87075836-5cb8-4014-b7f9-04c16c3e9e49" path="/var/lib/kubelet/pods/87075836-5cb8-4014-b7f9-04c16c3e9e49/volumes" Dec 01 16:03:21 crc kubenswrapper[4739]: I1201 16:03:21.307329 4739 scope.go:117] "RemoveContainer" containerID="d2b29e99374d12dfad1e0b93391e46de5fdf34979c0191fb2c48f9d1540bf8cc" Dec 01 16:03:21 crc kubenswrapper[4739]: I1201 16:03:21.350497 4739 scope.go:117] "RemoveContainer" containerID="d1b6b216abfc000d4755183b3e4f3b04155939bb36df24644d4bf8bdf1abfaa6" Dec 01 16:03:21 crc kubenswrapper[4739]: I1201 16:03:21.442197 4739 scope.go:117] "RemoveContainer" containerID="97509525b652865b0169cfe3b320f3b871a74aae156451e1bbb456d5290ffaf4" Dec 01 16:03:21 crc kubenswrapper[4739]: I1201 16:03:21.528355 4739 scope.go:117] "RemoveContainer" containerID="4acfd69e38cb39550e4d8ff81f3d3ff754b4f0442f2da03146a1e33cdee0c908" Dec 01 16:03:21 crc kubenswrapper[4739]: I1201 16:03:21.575058 4739 scope.go:117] "RemoveContainer" containerID="4cc201fe0207deb7ae939f6957d1ab463f42af0ddc04536d6010041aaecbb571" Dec 01 16:03:21 crc kubenswrapper[4739]: I1201 16:03:21.622716 4739 scope.go:117] "RemoveContainer" containerID="082718586a23ee88e13b30ddd5d473ead6afadad04225112b281f7207d54fc7d" Dec 01 16:03:21 crc kubenswrapper[4739]: I1201 16:03:21.638481 4739 scope.go:117] "RemoveContainer" containerID="d47ca4d362eedf2557a6ed7480efc0d503391d899c5e45e9a9b07a5e97240459" Dec 01 16:03:21 crc kubenswrapper[4739]: I1201 16:03:21.654015 4739 scope.go:117] "RemoveContainer" containerID="7babd0fe406838f7127a2d25735ec535eb730e9836f364ecba8cf00ca14eec41" Dec 01 16:03:21 crc kubenswrapper[4739]: I1201 16:03:21.677217 4739 scope.go:117] "RemoveContainer" containerID="1d957ff4235e37c97e3da8c855c3964759b82846beb116b74642e3259af6fb86" Dec 01 16:03:21 crc kubenswrapper[4739]: I1201 16:03:21.731174 4739 scope.go:117] "RemoveContainer" containerID="02d8fb088f94bd4c3c6e53a76b89db02aa2824ba0200335f87ab68cf963628a8" Dec 01 16:03:21 crc kubenswrapper[4739]: I1201 16:03:21.751000 4739 scope.go:117] "RemoveContainer" containerID="49e681dde4cd412b841676f21d70365bdcf6429930e93e74191ebbe9818216f1" Dec 01 16:03:26 crc kubenswrapper[4739]: I1201 16:03:26.477542 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:03:26 crc kubenswrapper[4739]: E1201 16:03:26.478822 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:03:29 crc kubenswrapper[4739]: I1201 16:03:29.034391 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-kfsn8"] Dec 01 16:03:29 crc kubenswrapper[4739]: I1201 16:03:29.048868 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-kfsn8"] Dec 01 16:03:30 crc kubenswrapper[4739]: I1201 16:03:30.492396 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cd4721c-4dfb-4219-9c69-c46299409e2a" path="/var/lib/kubelet/pods/0cd4721c-4dfb-4219-9c69-c46299409e2a/volumes" Dec 01 16:03:31 crc kubenswrapper[4739]: I1201 16:03:31.039142 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-g9q4b"] Dec 01 16:03:31 crc kubenswrapper[4739]: I1201 16:03:31.052630 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-g9q4b"] Dec 01 16:03:32 crc kubenswrapper[4739]: I1201 16:03:32.498656 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="588fd2a1-4b85-4af6-8b48-eff7fdf729ba" path="/var/lib/kubelet/pods/588fd2a1-4b85-4af6-8b48-eff7fdf729ba/volumes" Dec 01 16:03:40 crc kubenswrapper[4739]: I1201 16:03:40.477095 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:03:40 crc kubenswrapper[4739]: E1201 16:03:40.478220 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:03:53 crc kubenswrapper[4739]: I1201 16:03:53.477793 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:03:53 crc kubenswrapper[4739]: E1201 16:03:53.479031 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:03:56 crc kubenswrapper[4739]: I1201 16:03:56.119304 4739 generic.go:334] "Generic (PLEG): container finished" podID="ed881c2f-2f6c-40d1-85dd-42c2a9393fc9" containerID="2d1187e9457eac46edf1a1c390e3a91f1575ec4d78e403ede67f59bc9d257eb6" exitCode=0 Dec 01 16:03:56 crc kubenswrapper[4739]: I1201 16:03:56.119450 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" event={"ID":"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9","Type":"ContainerDied","Data":"2d1187e9457eac46edf1a1c390e3a91f1575ec4d78e403ede67f59bc9d257eb6"} Dec 01 16:03:57 crc kubenswrapper[4739]: I1201 16:03:57.518461 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" Dec 01 16:03:57 crc kubenswrapper[4739]: I1201 16:03:57.588308 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-inventory\") pod \"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9\" (UID: \"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9\") " Dec 01 16:03:57 crc kubenswrapper[4739]: I1201 16:03:57.588405 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-ssh-key\") pod \"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9\" (UID: \"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9\") " Dec 01 16:03:57 crc kubenswrapper[4739]: I1201 16:03:57.588682 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8qvk\" (UniqueName: \"kubernetes.io/projected/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-kube-api-access-f8qvk\") pod \"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9\" (UID: \"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9\") " Dec 01 16:03:57 crc kubenswrapper[4739]: I1201 16:03:57.597754 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-kube-api-access-f8qvk" (OuterVolumeSpecName: "kube-api-access-f8qvk") pod "ed881c2f-2f6c-40d1-85dd-42c2a9393fc9" (UID: "ed881c2f-2f6c-40d1-85dd-42c2a9393fc9"). InnerVolumeSpecName "kube-api-access-f8qvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:03:57 crc kubenswrapper[4739]: I1201 16:03:57.617360 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ed881c2f-2f6c-40d1-85dd-42c2a9393fc9" (UID: "ed881c2f-2f6c-40d1-85dd-42c2a9393fc9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:03:57 crc kubenswrapper[4739]: I1201 16:03:57.621899 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-inventory" (OuterVolumeSpecName: "inventory") pod "ed881c2f-2f6c-40d1-85dd-42c2a9393fc9" (UID: "ed881c2f-2f6c-40d1-85dd-42c2a9393fc9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:03:57 crc kubenswrapper[4739]: I1201 16:03:57.691447 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8qvk\" (UniqueName: \"kubernetes.io/projected/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-kube-api-access-f8qvk\") on node \"crc\" DevicePath \"\"" Dec 01 16:03:57 crc kubenswrapper[4739]: I1201 16:03:57.691493 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:03:57 crc kubenswrapper[4739]: I1201 16:03:57.691508 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.141008 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" event={"ID":"ed881c2f-2f6c-40d1-85dd-42c2a9393fc9","Type":"ContainerDied","Data":"81fd51f9bcc1443e8844b01ca99bfa3c64c39f60e67c5727ab877f5f09a48ce3"} Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.141299 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81fd51f9bcc1443e8844b01ca99bfa3c64c39f60e67c5727ab877f5f09a48ce3" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.141373 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.253205 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-7nr49"] Dec 01 16:03:58 crc kubenswrapper[4739]: E1201 16:03:58.253892 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed881c2f-2f6c-40d1-85dd-42c2a9393fc9" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.253937 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed881c2f-2f6c-40d1-85dd-42c2a9393fc9" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.254467 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed881c2f-2f6c-40d1-85dd-42c2a9393fc9" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.255766 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.260089 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.260125 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.260666 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.261857 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.264263 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-7nr49"] Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.303130 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/11851b7f-a7c3-462a-882e-a17612624b54-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-7nr49\" (UID: \"11851b7f-a7c3-462a-882e-a17612624b54\") " pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.303196 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z699t\" (UniqueName: \"kubernetes.io/projected/11851b7f-a7c3-462a-882e-a17612624b54-kube-api-access-z699t\") pod \"ssh-known-hosts-edpm-deployment-7nr49\" (UID: \"11851b7f-a7c3-462a-882e-a17612624b54\") " pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.303347 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/11851b7f-a7c3-462a-882e-a17612624b54-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-7nr49\" (UID: \"11851b7f-a7c3-462a-882e-a17612624b54\") " pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.404893 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z699t\" (UniqueName: \"kubernetes.io/projected/11851b7f-a7c3-462a-882e-a17612624b54-kube-api-access-z699t\") pod \"ssh-known-hosts-edpm-deployment-7nr49\" (UID: \"11851b7f-a7c3-462a-882e-a17612624b54\") " pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.404968 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/11851b7f-a7c3-462a-882e-a17612624b54-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-7nr49\" (UID: \"11851b7f-a7c3-462a-882e-a17612624b54\") " pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.405056 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/11851b7f-a7c3-462a-882e-a17612624b54-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-7nr49\" (UID: \"11851b7f-a7c3-462a-882e-a17612624b54\") " pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.410730 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/11851b7f-a7c3-462a-882e-a17612624b54-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-7nr49\" (UID: \"11851b7f-a7c3-462a-882e-a17612624b54\") " pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.412182 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/11851b7f-a7c3-462a-882e-a17612624b54-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-7nr49\" (UID: \"11851b7f-a7c3-462a-882e-a17612624b54\") " pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.424866 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z699t\" (UniqueName: \"kubernetes.io/projected/11851b7f-a7c3-462a-882e-a17612624b54-kube-api-access-z699t\") pod \"ssh-known-hosts-edpm-deployment-7nr49\" (UID: \"11851b7f-a7c3-462a-882e-a17612624b54\") " pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.596160 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:03:58 crc kubenswrapper[4739]: I1201 16:03:58.603682 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" Dec 01 16:03:59 crc kubenswrapper[4739]: I1201 16:03:59.161855 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-7nr49"] Dec 01 16:03:59 crc kubenswrapper[4739]: I1201 16:03:59.766927 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:04:00 crc kubenswrapper[4739]: I1201 16:04:00.169824 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" event={"ID":"11851b7f-a7c3-462a-882e-a17612624b54","Type":"ContainerStarted","Data":"01c05052ed43ba4db45fb5369fabc3240bf79c40f3bbe78d9d3dd947bc318eb3"} Dec 01 16:04:00 crc kubenswrapper[4739]: I1201 16:04:00.170168 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" event={"ID":"11851b7f-a7c3-462a-882e-a17612624b54","Type":"ContainerStarted","Data":"fec38ceec7abb2c091aad557ec2e81ab8239130cdff533db3220256bfd11adb1"} Dec 01 16:04:00 crc kubenswrapper[4739]: I1201 16:04:00.198467 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" podStartSLOduration=1.600478879 podStartE2EDuration="2.198444698s" podCreationTimestamp="2025-12-01 16:03:58 +0000 UTC" firstStartedPulling="2025-12-01 16:03:59.16514799 +0000 UTC m=+1740.990894084" lastFinishedPulling="2025-12-01 16:03:59.763113799 +0000 UTC m=+1741.588859903" observedRunningTime="2025-12-01 16:04:00.183835006 +0000 UTC m=+1742.009581100" watchObservedRunningTime="2025-12-01 16:04:00.198444698 +0000 UTC m=+1742.024190792" Dec 01 16:04:04 crc kubenswrapper[4739]: I1201 16:04:04.477379 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:04:04 crc kubenswrapper[4739]: E1201 16:04:04.477976 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:04:07 crc kubenswrapper[4739]: I1201 16:04:07.233668 4739 generic.go:334] "Generic (PLEG): container finished" podID="11851b7f-a7c3-462a-882e-a17612624b54" containerID="01c05052ed43ba4db45fb5369fabc3240bf79c40f3bbe78d9d3dd947bc318eb3" exitCode=0 Dec 01 16:04:07 crc kubenswrapper[4739]: I1201 16:04:07.233764 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" event={"ID":"11851b7f-a7c3-462a-882e-a17612624b54","Type":"ContainerDied","Data":"01c05052ed43ba4db45fb5369fabc3240bf79c40f3bbe78d9d3dd947bc318eb3"} Dec 01 16:04:08 crc kubenswrapper[4739]: I1201 16:04:08.747966 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" Dec 01 16:04:08 crc kubenswrapper[4739]: I1201 16:04:08.799263 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/11851b7f-a7c3-462a-882e-a17612624b54-ssh-key-openstack-edpm-ipam\") pod \"11851b7f-a7c3-462a-882e-a17612624b54\" (UID: \"11851b7f-a7c3-462a-882e-a17612624b54\") " Dec 01 16:04:08 crc kubenswrapper[4739]: I1201 16:04:08.799350 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/11851b7f-a7c3-462a-882e-a17612624b54-inventory-0\") pod \"11851b7f-a7c3-462a-882e-a17612624b54\" (UID: \"11851b7f-a7c3-462a-882e-a17612624b54\") " Dec 01 16:04:08 crc kubenswrapper[4739]: I1201 16:04:08.799465 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z699t\" (UniqueName: \"kubernetes.io/projected/11851b7f-a7c3-462a-882e-a17612624b54-kube-api-access-z699t\") pod \"11851b7f-a7c3-462a-882e-a17612624b54\" (UID: \"11851b7f-a7c3-462a-882e-a17612624b54\") " Dec 01 16:04:08 crc kubenswrapper[4739]: I1201 16:04:08.807055 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11851b7f-a7c3-462a-882e-a17612624b54-kube-api-access-z699t" (OuterVolumeSpecName: "kube-api-access-z699t") pod "11851b7f-a7c3-462a-882e-a17612624b54" (UID: "11851b7f-a7c3-462a-882e-a17612624b54"). InnerVolumeSpecName "kube-api-access-z699t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:04:08 crc kubenswrapper[4739]: I1201 16:04:08.833593 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11851b7f-a7c3-462a-882e-a17612624b54-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "11851b7f-a7c3-462a-882e-a17612624b54" (UID: "11851b7f-a7c3-462a-882e-a17612624b54"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:04:08 crc kubenswrapper[4739]: I1201 16:04:08.838904 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11851b7f-a7c3-462a-882e-a17612624b54-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "11851b7f-a7c3-462a-882e-a17612624b54" (UID: "11851b7f-a7c3-462a-882e-a17612624b54"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:04:08 crc kubenswrapper[4739]: I1201 16:04:08.901532 4739 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/11851b7f-a7c3-462a-882e-a17612624b54-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:04:08 crc kubenswrapper[4739]: I1201 16:04:08.901563 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z699t\" (UniqueName: \"kubernetes.io/projected/11851b7f-a7c3-462a-882e-a17612624b54-kube-api-access-z699t\") on node \"crc\" DevicePath \"\"" Dec 01 16:04:08 crc kubenswrapper[4739]: I1201 16:04:08.901573 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/11851b7f-a7c3-462a-882e-a17612624b54-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.261203 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" event={"ID":"11851b7f-a7c3-462a-882e-a17612624b54","Type":"ContainerDied","Data":"fec38ceec7abb2c091aad557ec2e81ab8239130cdff533db3220256bfd11adb1"} Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.261783 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fec38ceec7abb2c091aad557ec2e81ab8239130cdff533db3220256bfd11adb1" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.261346 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-7nr49" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.365114 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c"] Dec 01 16:04:09 crc kubenswrapper[4739]: E1201 16:04:09.365636 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11851b7f-a7c3-462a-882e-a17612624b54" containerName="ssh-known-hosts-edpm-deployment" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.365658 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="11851b7f-a7c3-462a-882e-a17612624b54" containerName="ssh-known-hosts-edpm-deployment" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.365864 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="11851b7f-a7c3-462a-882e-a17612624b54" containerName="ssh-known-hosts-edpm-deployment" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.366573 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.376223 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c"] Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.403637 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.403978 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.404597 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.404937 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.410282 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/750dd9de-c593-43c1-845f-50971d0fdb6e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vn45c\" (UID: \"750dd9de-c593-43c1-845f-50971d0fdb6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.410724 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79gbn\" (UniqueName: \"kubernetes.io/projected/750dd9de-c593-43c1-845f-50971d0fdb6e-kube-api-access-79gbn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vn45c\" (UID: \"750dd9de-c593-43c1-845f-50971d0fdb6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.410914 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/750dd9de-c593-43c1-845f-50971d0fdb6e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vn45c\" (UID: \"750dd9de-c593-43c1-845f-50971d0fdb6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.512791 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/750dd9de-c593-43c1-845f-50971d0fdb6e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vn45c\" (UID: \"750dd9de-c593-43c1-845f-50971d0fdb6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.512885 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/750dd9de-c593-43c1-845f-50971d0fdb6e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vn45c\" (UID: \"750dd9de-c593-43c1-845f-50971d0fdb6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.513068 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79gbn\" (UniqueName: \"kubernetes.io/projected/750dd9de-c593-43c1-845f-50971d0fdb6e-kube-api-access-79gbn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vn45c\" (UID: \"750dd9de-c593-43c1-845f-50971d0fdb6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.518295 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/750dd9de-c593-43c1-845f-50971d0fdb6e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vn45c\" (UID: \"750dd9de-c593-43c1-845f-50971d0fdb6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.518390 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/750dd9de-c593-43c1-845f-50971d0fdb6e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vn45c\" (UID: \"750dd9de-c593-43c1-845f-50971d0fdb6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.542287 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79gbn\" (UniqueName: \"kubernetes.io/projected/750dd9de-c593-43c1-845f-50971d0fdb6e-kube-api-access-79gbn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vn45c\" (UID: \"750dd9de-c593-43c1-845f-50971d0fdb6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" Dec 01 16:04:09 crc kubenswrapper[4739]: I1201 16:04:09.729295 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" Dec 01 16:04:10 crc kubenswrapper[4739]: I1201 16:04:10.314392 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c"] Dec 01 16:04:11 crc kubenswrapper[4739]: I1201 16:04:11.290576 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" event={"ID":"750dd9de-c593-43c1-845f-50971d0fdb6e","Type":"ContainerStarted","Data":"c19e5ee484b4d715374fa88aa963e7ff3ce05e47a1cf8f3e974706dcfd923a61"} Dec 01 16:04:12 crc kubenswrapper[4739]: I1201 16:04:12.299443 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" event={"ID":"750dd9de-c593-43c1-845f-50971d0fdb6e","Type":"ContainerStarted","Data":"7b6e73dee68461b7d90ae8d167ee91b2fed52c706164e5fafbe31e652dbb61c5"} Dec 01 16:04:12 crc kubenswrapper[4739]: I1201 16:04:12.317995 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" podStartSLOduration=1.987125475 podStartE2EDuration="3.317977413s" podCreationTimestamp="2025-12-01 16:04:09 +0000 UTC" firstStartedPulling="2025-12-01 16:04:10.301482756 +0000 UTC m=+1752.127228850" lastFinishedPulling="2025-12-01 16:04:11.632334694 +0000 UTC m=+1753.458080788" observedRunningTime="2025-12-01 16:04:12.313434493 +0000 UTC m=+1754.139180597" watchObservedRunningTime="2025-12-01 16:04:12.317977413 +0000 UTC m=+1754.143723507" Dec 01 16:04:16 crc kubenswrapper[4739]: I1201 16:04:16.059637 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-xtpmt"] Dec 01 16:04:16 crc kubenswrapper[4739]: I1201 16:04:16.082482 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-wfs6d"] Dec 01 16:04:16 crc kubenswrapper[4739]: I1201 16:04:16.090771 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-8rbwc"] Dec 01 16:04:16 crc kubenswrapper[4739]: I1201 16:04:16.109495 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-xtpmt"] Dec 01 16:04:16 crc kubenswrapper[4739]: I1201 16:04:16.121678 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-wfs6d"] Dec 01 16:04:16 crc kubenswrapper[4739]: I1201 16:04:16.131582 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-d474-account-create-update-g7qz8"] Dec 01 16:04:16 crc kubenswrapper[4739]: I1201 16:04:16.140291 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-d474-account-create-update-g7qz8"] Dec 01 16:04:16 crc kubenswrapper[4739]: I1201 16:04:16.147251 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-8rbwc"] Dec 01 16:04:16 crc kubenswrapper[4739]: I1201 16:04:16.489998 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15" path="/var/lib/kubelet/pods/2e9c92eb-e9cf-4b42-8b10-5e38e7fd6c15/volumes" Dec 01 16:04:16 crc kubenswrapper[4739]: I1201 16:04:16.491006 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d7f9b4c-d167-4b93-9a74-abb347111f76" path="/var/lib/kubelet/pods/4d7f9b4c-d167-4b93-9a74-abb347111f76/volumes" Dec 01 16:04:16 crc kubenswrapper[4739]: I1201 16:04:16.491882 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="594ff868-2963-45d0-ba4f-8e96f7e2afa1" path="/var/lib/kubelet/pods/594ff868-2963-45d0-ba4f-8e96f7e2afa1/volumes" Dec 01 16:04:16 crc kubenswrapper[4739]: I1201 16:04:16.492738 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c795dbe9-11c4-4ae2-a20f-63e964a8665c" path="/var/lib/kubelet/pods/c795dbe9-11c4-4ae2-a20f-63e964a8665c/volumes" Dec 01 16:04:17 crc kubenswrapper[4739]: I1201 16:04:17.033787 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-5638-account-create-update-tcl6s"] Dec 01 16:04:17 crc kubenswrapper[4739]: I1201 16:04:17.041313 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-8179-account-create-update-57dw7"] Dec 01 16:04:17 crc kubenswrapper[4739]: I1201 16:04:17.049683 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-5638-account-create-update-tcl6s"] Dec 01 16:04:17 crc kubenswrapper[4739]: I1201 16:04:17.058353 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-8179-account-create-update-57dw7"] Dec 01 16:04:17 crc kubenswrapper[4739]: I1201 16:04:17.478207 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:04:17 crc kubenswrapper[4739]: E1201 16:04:17.478688 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:04:18 crc kubenswrapper[4739]: I1201 16:04:18.493938 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aa61fba-e9af-4666-a96e-498c56ddeaf0" path="/var/lib/kubelet/pods/8aa61fba-e9af-4666-a96e-498c56ddeaf0/volumes" Dec 01 16:04:18 crc kubenswrapper[4739]: I1201 16:04:18.494813 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f400ecb1-7790-4325-a68d-a9ae8e8527d1" path="/var/lib/kubelet/pods/f400ecb1-7790-4325-a68d-a9ae8e8527d1/volumes" Dec 01 16:04:21 crc kubenswrapper[4739]: I1201 16:04:21.382935 4739 generic.go:334] "Generic (PLEG): container finished" podID="750dd9de-c593-43c1-845f-50971d0fdb6e" containerID="7b6e73dee68461b7d90ae8d167ee91b2fed52c706164e5fafbe31e652dbb61c5" exitCode=0 Dec 01 16:04:21 crc kubenswrapper[4739]: I1201 16:04:21.383039 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" event={"ID":"750dd9de-c593-43c1-845f-50971d0fdb6e","Type":"ContainerDied","Data":"7b6e73dee68461b7d90ae8d167ee91b2fed52c706164e5fafbe31e652dbb61c5"} Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.019509 4739 scope.go:117] "RemoveContainer" containerID="3f6ef0a436a192023393027b9011e2de3fbc772e6a45cf037ee55a136e52ab94" Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.058768 4739 scope.go:117] "RemoveContainer" containerID="8e76b1a4598841709a798c7d605f6b410086ae6f2e45802299e93ac52f40fe9d" Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.124115 4739 scope.go:117] "RemoveContainer" containerID="e9fe065410204039b095972e576064aef9d5b8a20668e67d7eaf2fa1ff8f426d" Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.156155 4739 scope.go:117] "RemoveContainer" containerID="67d173165ff5bf022fd97d4534a03846fdf75b53cc059c6e8843f3c61276236b" Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.197327 4739 scope.go:117] "RemoveContainer" containerID="0b19591959da1c430cf8a8d6003adce34472fd08a3f8f479563df60f8174d8bb" Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.238636 4739 scope.go:117] "RemoveContainer" containerID="09b40f88daf06ba6433f93b1e0b683e201d24da698cdacf14a17f83055f351a6" Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.273250 4739 scope.go:117] "RemoveContainer" containerID="1be628c6fd28e82ded158029d635c2a2e5ef07c2e4fee3bc7ad33c8fc016442d" Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.297586 4739 scope.go:117] "RemoveContainer" containerID="539633d585cc574c461499a29503013b1bac17403354c7b02502c66ae57990b4" Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.816364 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.864881 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/750dd9de-c593-43c1-845f-50971d0fdb6e-ssh-key\") pod \"750dd9de-c593-43c1-845f-50971d0fdb6e\" (UID: \"750dd9de-c593-43c1-845f-50971d0fdb6e\") " Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.865083 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79gbn\" (UniqueName: \"kubernetes.io/projected/750dd9de-c593-43c1-845f-50971d0fdb6e-kube-api-access-79gbn\") pod \"750dd9de-c593-43c1-845f-50971d0fdb6e\" (UID: \"750dd9de-c593-43c1-845f-50971d0fdb6e\") " Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.865128 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/750dd9de-c593-43c1-845f-50971d0fdb6e-inventory\") pod \"750dd9de-c593-43c1-845f-50971d0fdb6e\" (UID: \"750dd9de-c593-43c1-845f-50971d0fdb6e\") " Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.873803 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/750dd9de-c593-43c1-845f-50971d0fdb6e-kube-api-access-79gbn" (OuterVolumeSpecName: "kube-api-access-79gbn") pod "750dd9de-c593-43c1-845f-50971d0fdb6e" (UID: "750dd9de-c593-43c1-845f-50971d0fdb6e"). InnerVolumeSpecName "kube-api-access-79gbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:04:22 crc kubenswrapper[4739]: E1201 16:04:22.895815 4739 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/750dd9de-c593-43c1-845f-50971d0fdb6e-ssh-key podName:750dd9de-c593-43c1-845f-50971d0fdb6e nodeName:}" failed. No retries permitted until 2025-12-01 16:04:23.39577743 +0000 UTC m=+1765.221523594 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key" (UniqueName: "kubernetes.io/secret/750dd9de-c593-43c1-845f-50971d0fdb6e-ssh-key") pod "750dd9de-c593-43c1-845f-50971d0fdb6e" (UID: "750dd9de-c593-43c1-845f-50971d0fdb6e") : error deleting /var/lib/kubelet/pods/750dd9de-c593-43c1-845f-50971d0fdb6e/volume-subpaths: remove /var/lib/kubelet/pods/750dd9de-c593-43c1-845f-50971d0fdb6e/volume-subpaths: no such file or directory Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.899215 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/750dd9de-c593-43c1-845f-50971d0fdb6e-inventory" (OuterVolumeSpecName: "inventory") pod "750dd9de-c593-43c1-845f-50971d0fdb6e" (UID: "750dd9de-c593-43c1-845f-50971d0fdb6e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.967458 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79gbn\" (UniqueName: \"kubernetes.io/projected/750dd9de-c593-43c1-845f-50971d0fdb6e-kube-api-access-79gbn\") on node \"crc\" DevicePath \"\"" Dec 01 16:04:22 crc kubenswrapper[4739]: I1201 16:04:22.967529 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/750dd9de-c593-43c1-845f-50971d0fdb6e-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.411924 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" event={"ID":"750dd9de-c593-43c1-845f-50971d0fdb6e","Type":"ContainerDied","Data":"c19e5ee484b4d715374fa88aa963e7ff3ce05e47a1cf8f3e974706dcfd923a61"} Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.412237 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c19e5ee484b4d715374fa88aa963e7ff3ce05e47a1cf8f3e974706dcfd923a61" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.412160 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.477109 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/750dd9de-c593-43c1-845f-50971d0fdb6e-ssh-key\") pod \"750dd9de-c593-43c1-845f-50971d0fdb6e\" (UID: \"750dd9de-c593-43c1-845f-50971d0fdb6e\") " Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.479810 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d"] Dec 01 16:04:23 crc kubenswrapper[4739]: E1201 16:04:23.480293 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="750dd9de-c593-43c1-845f-50971d0fdb6e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.480595 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="750dd9de-c593-43c1-845f-50971d0fdb6e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.480928 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="750dd9de-c593-43c1-845f-50971d0fdb6e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.481685 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/750dd9de-c593-43c1-845f-50971d0fdb6e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "750dd9de-c593-43c1-845f-50971d0fdb6e" (UID: "750dd9de-c593-43c1-845f-50971d0fdb6e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.481793 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.503050 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d"] Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.578665 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ac2d686-8407-495e-b4e2-1989d8268492-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d\" (UID: \"6ac2d686-8407-495e-b4e2-1989d8268492\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.578738 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qj8s\" (UniqueName: \"kubernetes.io/projected/6ac2d686-8407-495e-b4e2-1989d8268492-kube-api-access-9qj8s\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d\" (UID: \"6ac2d686-8407-495e-b4e2-1989d8268492\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.578785 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ac2d686-8407-495e-b4e2-1989d8268492-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d\" (UID: \"6ac2d686-8407-495e-b4e2-1989d8268492\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.579210 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/750dd9de-c593-43c1-845f-50971d0fdb6e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.681259 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ac2d686-8407-495e-b4e2-1989d8268492-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d\" (UID: \"6ac2d686-8407-495e-b4e2-1989d8268492\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.681368 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qj8s\" (UniqueName: \"kubernetes.io/projected/6ac2d686-8407-495e-b4e2-1989d8268492-kube-api-access-9qj8s\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d\" (UID: \"6ac2d686-8407-495e-b4e2-1989d8268492\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.681449 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ac2d686-8407-495e-b4e2-1989d8268492-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d\" (UID: \"6ac2d686-8407-495e-b4e2-1989d8268492\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.687068 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ac2d686-8407-495e-b4e2-1989d8268492-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d\" (UID: \"6ac2d686-8407-495e-b4e2-1989d8268492\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.687685 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ac2d686-8407-495e-b4e2-1989d8268492-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d\" (UID: \"6ac2d686-8407-495e-b4e2-1989d8268492\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.704411 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qj8s\" (UniqueName: \"kubernetes.io/projected/6ac2d686-8407-495e-b4e2-1989d8268492-kube-api-access-9qj8s\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d\" (UID: \"6ac2d686-8407-495e-b4e2-1989d8268492\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" Dec 01 16:04:23 crc kubenswrapper[4739]: I1201 16:04:23.844186 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" Dec 01 16:04:24 crc kubenswrapper[4739]: I1201 16:04:24.389682 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d"] Dec 01 16:04:24 crc kubenswrapper[4739]: I1201 16:04:24.422550 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" event={"ID":"6ac2d686-8407-495e-b4e2-1989d8268492","Type":"ContainerStarted","Data":"c0a0ffc244e5cecf60e4b13bafb4357f56462caa22d8013724f0553f0635ac8a"} Dec 01 16:04:25 crc kubenswrapper[4739]: I1201 16:04:25.436405 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" event={"ID":"6ac2d686-8407-495e-b4e2-1989d8268492","Type":"ContainerStarted","Data":"b3bb33d102750bb8da2d403c3a0abe849107eede5f0b0b87d7010ea788e82eb2"} Dec 01 16:04:25 crc kubenswrapper[4739]: I1201 16:04:25.462248 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" podStartSLOduration=1.768589924 podStartE2EDuration="2.46222736s" podCreationTimestamp="2025-12-01 16:04:23 +0000 UTC" firstStartedPulling="2025-12-01 16:04:24.39651882 +0000 UTC m=+1766.222264924" lastFinishedPulling="2025-12-01 16:04:25.090156266 +0000 UTC m=+1766.915902360" observedRunningTime="2025-12-01 16:04:25.452629372 +0000 UTC m=+1767.278375486" watchObservedRunningTime="2025-12-01 16:04:25.46222736 +0000 UTC m=+1767.287973454" Dec 01 16:04:29 crc kubenswrapper[4739]: I1201 16:04:29.477897 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:04:29 crc kubenswrapper[4739]: E1201 16:04:29.480771 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:04:38 crc kubenswrapper[4739]: I1201 16:04:38.046478 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-626sv"] Dec 01 16:04:38 crc kubenswrapper[4739]: I1201 16:04:38.054308 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-626sv"] Dec 01 16:04:38 crc kubenswrapper[4739]: I1201 16:04:38.487580 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dcf654f-3264-479d-9d65-acd256667075" path="/var/lib/kubelet/pods/4dcf654f-3264-479d-9d65-acd256667075/volumes" Dec 01 16:04:38 crc kubenswrapper[4739]: I1201 16:04:38.579529 4739 generic.go:334] "Generic (PLEG): container finished" podID="6ac2d686-8407-495e-b4e2-1989d8268492" containerID="b3bb33d102750bb8da2d403c3a0abe849107eede5f0b0b87d7010ea788e82eb2" exitCode=0 Dec 01 16:04:38 crc kubenswrapper[4739]: I1201 16:04:38.579577 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" event={"ID":"6ac2d686-8407-495e-b4e2-1989d8268492","Type":"ContainerDied","Data":"b3bb33d102750bb8da2d403c3a0abe849107eede5f0b0b87d7010ea788e82eb2"} Dec 01 16:04:39 crc kubenswrapper[4739]: I1201 16:04:39.962697 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" Dec 01 16:04:40 crc kubenswrapper[4739]: I1201 16:04:40.122208 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ac2d686-8407-495e-b4e2-1989d8268492-inventory\") pod \"6ac2d686-8407-495e-b4e2-1989d8268492\" (UID: \"6ac2d686-8407-495e-b4e2-1989d8268492\") " Dec 01 16:04:40 crc kubenswrapper[4739]: I1201 16:04:40.122388 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ac2d686-8407-495e-b4e2-1989d8268492-ssh-key\") pod \"6ac2d686-8407-495e-b4e2-1989d8268492\" (UID: \"6ac2d686-8407-495e-b4e2-1989d8268492\") " Dec 01 16:04:40 crc kubenswrapper[4739]: I1201 16:04:40.122452 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qj8s\" (UniqueName: \"kubernetes.io/projected/6ac2d686-8407-495e-b4e2-1989d8268492-kube-api-access-9qj8s\") pod \"6ac2d686-8407-495e-b4e2-1989d8268492\" (UID: \"6ac2d686-8407-495e-b4e2-1989d8268492\") " Dec 01 16:04:40 crc kubenswrapper[4739]: I1201 16:04:40.129159 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ac2d686-8407-495e-b4e2-1989d8268492-kube-api-access-9qj8s" (OuterVolumeSpecName: "kube-api-access-9qj8s") pod "6ac2d686-8407-495e-b4e2-1989d8268492" (UID: "6ac2d686-8407-495e-b4e2-1989d8268492"). InnerVolumeSpecName "kube-api-access-9qj8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:04:40 crc kubenswrapper[4739]: I1201 16:04:40.153353 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ac2d686-8407-495e-b4e2-1989d8268492-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6ac2d686-8407-495e-b4e2-1989d8268492" (UID: "6ac2d686-8407-495e-b4e2-1989d8268492"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:04:40 crc kubenswrapper[4739]: I1201 16:04:40.170161 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ac2d686-8407-495e-b4e2-1989d8268492-inventory" (OuterVolumeSpecName: "inventory") pod "6ac2d686-8407-495e-b4e2-1989d8268492" (UID: "6ac2d686-8407-495e-b4e2-1989d8268492"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:04:40 crc kubenswrapper[4739]: I1201 16:04:40.224061 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ac2d686-8407-495e-b4e2-1989d8268492-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:04:40 crc kubenswrapper[4739]: I1201 16:04:40.224108 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qj8s\" (UniqueName: \"kubernetes.io/projected/6ac2d686-8407-495e-b4e2-1989d8268492-kube-api-access-9qj8s\") on node \"crc\" DevicePath \"\"" Dec 01 16:04:40 crc kubenswrapper[4739]: I1201 16:04:40.224124 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ac2d686-8407-495e-b4e2-1989d8268492-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:04:40 crc kubenswrapper[4739]: I1201 16:04:40.596627 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" event={"ID":"6ac2d686-8407-495e-b4e2-1989d8268492","Type":"ContainerDied","Data":"c0a0ffc244e5cecf60e4b13bafb4357f56462caa22d8013724f0553f0635ac8a"} Dec 01 16:04:40 crc kubenswrapper[4739]: I1201 16:04:40.596678 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0a0ffc244e5cecf60e4b13bafb4357f56462caa22d8013724f0553f0635ac8a" Dec 01 16:04:40 crc kubenswrapper[4739]: I1201 16:04:40.596685 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d" Dec 01 16:04:43 crc kubenswrapper[4739]: I1201 16:04:43.478862 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:04:44 crc kubenswrapper[4739]: I1201 16:04:44.642407 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"ddde4c9dc4ce14363e987196bc7779dbb2cfefa0de2f710f51b5795ef1b081bd"} Dec 01 16:05:00 crc kubenswrapper[4739]: I1201 16:05:00.044835 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-clm6v"] Dec 01 16:05:00 crc kubenswrapper[4739]: I1201 16:05:00.059446 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-clm6v"] Dec 01 16:05:00 crc kubenswrapper[4739]: I1201 16:05:00.067971 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-t9wxt"] Dec 01 16:05:00 crc kubenswrapper[4739]: I1201 16:05:00.076304 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-t9wxt"] Dec 01 16:05:00 crc kubenswrapper[4739]: I1201 16:05:00.496935 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1af0156-b7e0-4257-a1ad-2ae110ae8d54" path="/var/lib/kubelet/pods/a1af0156-b7e0-4257-a1ad-2ae110ae8d54/volumes" Dec 01 16:05:00 crc kubenswrapper[4739]: I1201 16:05:00.498305 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3b310c6-127c-45ba-be81-bb6cd95fcb7b" path="/var/lib/kubelet/pods/e3b310c6-127c-45ba-be81-bb6cd95fcb7b/volumes" Dec 01 16:05:22 crc kubenswrapper[4739]: I1201 16:05:22.511196 4739 scope.go:117] "RemoveContainer" containerID="19165388877e6a8140fc0e43fbaa05e81a77f89bed4844ade58d6f86738dc28c" Dec 01 16:05:22 crc kubenswrapper[4739]: I1201 16:05:22.572714 4739 scope.go:117] "RemoveContainer" containerID="6f84adc8bc9432bd88b2da8732675bf16b677ebba7e9df93da652eeffb56da30" Dec 01 16:05:22 crc kubenswrapper[4739]: I1201 16:05:22.632290 4739 scope.go:117] "RemoveContainer" containerID="919e6ab567e8b3ac14962fa8ab6d6a986cd9e8f3d3b1db1c46dc5db76295131c" Dec 01 16:05:45 crc kubenswrapper[4739]: I1201 16:05:45.050129 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-sh566"] Dec 01 16:05:45 crc kubenswrapper[4739]: I1201 16:05:45.057038 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-sh566"] Dec 01 16:05:46 crc kubenswrapper[4739]: I1201 16:05:46.486975 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="951ace80-f656-4e64-85e7-d4d54d793159" path="/var/lib/kubelet/pods/951ace80-f656-4e64-85e7-d4d54d793159/volumes" Dec 01 16:06:22 crc kubenswrapper[4739]: I1201 16:06:22.745205 4739 scope.go:117] "RemoveContainer" containerID="9bf623b03225199969eac35676c02ac9a56eddd549610375275b661f7308174e" Dec 01 16:07:09 crc kubenswrapper[4739]: I1201 16:07:09.622330 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:07:09 crc kubenswrapper[4739]: I1201 16:07:09.623144 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:07:39 crc kubenswrapper[4739]: I1201 16:07:39.622479 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:07:39 crc kubenswrapper[4739]: I1201 16:07:39.623093 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:08:09 crc kubenswrapper[4739]: I1201 16:08:09.622041 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:08:09 crc kubenswrapper[4739]: I1201 16:08:09.622987 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:08:09 crc kubenswrapper[4739]: I1201 16:08:09.623062 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 16:08:09 crc kubenswrapper[4739]: I1201 16:08:09.624037 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ddde4c9dc4ce14363e987196bc7779dbb2cfefa0de2f710f51b5795ef1b081bd"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:08:09 crc kubenswrapper[4739]: I1201 16:08:09.624116 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://ddde4c9dc4ce14363e987196bc7779dbb2cfefa0de2f710f51b5795ef1b081bd" gracePeriod=600 Dec 01 16:08:10 crc kubenswrapper[4739]: I1201 16:08:10.172149 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="ddde4c9dc4ce14363e987196bc7779dbb2cfefa0de2f710f51b5795ef1b081bd" exitCode=0 Dec 01 16:08:10 crc kubenswrapper[4739]: I1201 16:08:10.172258 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"ddde4c9dc4ce14363e987196bc7779dbb2cfefa0de2f710f51b5795ef1b081bd"} Dec 01 16:08:10 crc kubenswrapper[4739]: I1201 16:08:10.172779 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be"} Dec 01 16:08:10 crc kubenswrapper[4739]: I1201 16:08:10.172855 4739 scope.go:117] "RemoveContainer" containerID="9582b96c165a86459310130bd83347d7788248a752b952322575e21486ae639a" Dec 01 16:08:56 crc kubenswrapper[4739]: I1201 16:08:56.898607 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-84g49"] Dec 01 16:08:56 crc kubenswrapper[4739]: E1201 16:08:56.899907 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac2d686-8407-495e-b4e2-1989d8268492" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:08:56 crc kubenswrapper[4739]: I1201 16:08:56.899930 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac2d686-8407-495e-b4e2-1989d8268492" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:08:56 crc kubenswrapper[4739]: I1201 16:08:56.900182 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ac2d686-8407-495e-b4e2-1989d8268492" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:08:56 crc kubenswrapper[4739]: I1201 16:08:56.902241 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:08:56 crc kubenswrapper[4739]: I1201 16:08:56.913548 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-84g49"] Dec 01 16:08:57 crc kubenswrapper[4739]: I1201 16:08:57.013495 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac37c1aa-2809-4754-9373-fed7588f548e-utilities\") pod \"redhat-operators-84g49\" (UID: \"ac37c1aa-2809-4754-9373-fed7588f548e\") " pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:08:57 crc kubenswrapper[4739]: I1201 16:08:57.013557 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac37c1aa-2809-4754-9373-fed7588f548e-catalog-content\") pod \"redhat-operators-84g49\" (UID: \"ac37c1aa-2809-4754-9373-fed7588f548e\") " pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:08:57 crc kubenswrapper[4739]: I1201 16:08:57.013603 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbntn\" (UniqueName: \"kubernetes.io/projected/ac37c1aa-2809-4754-9373-fed7588f548e-kube-api-access-xbntn\") pod \"redhat-operators-84g49\" (UID: \"ac37c1aa-2809-4754-9373-fed7588f548e\") " pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:08:57 crc kubenswrapper[4739]: I1201 16:08:57.115915 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac37c1aa-2809-4754-9373-fed7588f548e-utilities\") pod \"redhat-operators-84g49\" (UID: \"ac37c1aa-2809-4754-9373-fed7588f548e\") " pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:08:57 crc kubenswrapper[4739]: I1201 16:08:57.115988 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac37c1aa-2809-4754-9373-fed7588f548e-catalog-content\") pod \"redhat-operators-84g49\" (UID: \"ac37c1aa-2809-4754-9373-fed7588f548e\") " pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:08:57 crc kubenswrapper[4739]: I1201 16:08:57.116045 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbntn\" (UniqueName: \"kubernetes.io/projected/ac37c1aa-2809-4754-9373-fed7588f548e-kube-api-access-xbntn\") pod \"redhat-operators-84g49\" (UID: \"ac37c1aa-2809-4754-9373-fed7588f548e\") " pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:08:57 crc kubenswrapper[4739]: I1201 16:08:57.117498 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac37c1aa-2809-4754-9373-fed7588f548e-utilities\") pod \"redhat-operators-84g49\" (UID: \"ac37c1aa-2809-4754-9373-fed7588f548e\") " pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:08:57 crc kubenswrapper[4739]: I1201 16:08:57.117686 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac37c1aa-2809-4754-9373-fed7588f548e-catalog-content\") pod \"redhat-operators-84g49\" (UID: \"ac37c1aa-2809-4754-9373-fed7588f548e\") " pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:08:57 crc kubenswrapper[4739]: I1201 16:08:57.140195 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbntn\" (UniqueName: \"kubernetes.io/projected/ac37c1aa-2809-4754-9373-fed7588f548e-kube-api-access-xbntn\") pod \"redhat-operators-84g49\" (UID: \"ac37c1aa-2809-4754-9373-fed7588f548e\") " pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:08:57 crc kubenswrapper[4739]: I1201 16:08:57.235538 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:08:57 crc kubenswrapper[4739]: I1201 16:08:57.744186 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-84g49"] Dec 01 16:08:58 crc kubenswrapper[4739]: I1201 16:08:58.750375 4739 generic.go:334] "Generic (PLEG): container finished" podID="ac37c1aa-2809-4754-9373-fed7588f548e" containerID="b03b03ed6b69cae86ec843094ee3438d62f64c36d5234512924cd1b5125a9564" exitCode=0 Dec 01 16:08:58 crc kubenswrapper[4739]: I1201 16:08:58.750477 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84g49" event={"ID":"ac37c1aa-2809-4754-9373-fed7588f548e","Type":"ContainerDied","Data":"b03b03ed6b69cae86ec843094ee3438d62f64c36d5234512924cd1b5125a9564"} Dec 01 16:08:58 crc kubenswrapper[4739]: I1201 16:08:58.751328 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84g49" event={"ID":"ac37c1aa-2809-4754-9373-fed7588f548e","Type":"ContainerStarted","Data":"f45e1e856c83b43eeadb6a9400d2234b627613312b66ba9a1f15cfde189690a8"} Dec 01 16:08:58 crc kubenswrapper[4739]: I1201 16:08:58.752786 4739 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:09:00 crc kubenswrapper[4739]: I1201 16:09:00.771812 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84g49" event={"ID":"ac37c1aa-2809-4754-9373-fed7588f548e","Type":"ContainerStarted","Data":"c535241e2fff426d223e6b4c5ed481676d6329b6b70840614b693541b209ca41"} Dec 01 16:09:02 crc kubenswrapper[4739]: I1201 16:09:02.796499 4739 generic.go:334] "Generic (PLEG): container finished" podID="ac37c1aa-2809-4754-9373-fed7588f548e" containerID="c535241e2fff426d223e6b4c5ed481676d6329b6b70840614b693541b209ca41" exitCode=0 Dec 01 16:09:02 crc kubenswrapper[4739]: I1201 16:09:02.796567 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84g49" event={"ID":"ac37c1aa-2809-4754-9373-fed7588f548e","Type":"ContainerDied","Data":"c535241e2fff426d223e6b4c5ed481676d6329b6b70840614b693541b209ca41"} Dec 01 16:09:03 crc kubenswrapper[4739]: I1201 16:09:03.818571 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84g49" event={"ID":"ac37c1aa-2809-4754-9373-fed7588f548e","Type":"ContainerStarted","Data":"2b478dc2fac0dbf971ab0caff4cb683a41137faeba0cec87d8797caba652ea40"} Dec 01 16:09:03 crc kubenswrapper[4739]: I1201 16:09:03.846612 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-84g49" podStartSLOduration=3.319309601 podStartE2EDuration="7.846590541s" podCreationTimestamp="2025-12-01 16:08:56 +0000 UTC" firstStartedPulling="2025-12-01 16:08:58.752508944 +0000 UTC m=+2040.578255038" lastFinishedPulling="2025-12-01 16:09:03.279789884 +0000 UTC m=+2045.105535978" observedRunningTime="2025-12-01 16:09:03.837317704 +0000 UTC m=+2045.663063808" watchObservedRunningTime="2025-12-01 16:09:03.846590541 +0000 UTC m=+2045.672336635" Dec 01 16:09:07 crc kubenswrapper[4739]: I1201 16:09:07.236889 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:09:07 crc kubenswrapper[4739]: I1201 16:09:07.237487 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:09:08 crc kubenswrapper[4739]: I1201 16:09:08.309671 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-84g49" podUID="ac37c1aa-2809-4754-9373-fed7588f548e" containerName="registry-server" probeResult="failure" output=< Dec 01 16:09:08 crc kubenswrapper[4739]: timeout: failed to connect service ":50051" within 1s Dec 01 16:09:08 crc kubenswrapper[4739]: > Dec 01 16:09:17 crc kubenswrapper[4739]: I1201 16:09:17.325293 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:09:17 crc kubenswrapper[4739]: I1201 16:09:17.408611 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:09:17 crc kubenswrapper[4739]: I1201 16:09:17.573359 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-84g49"] Dec 01 16:09:18 crc kubenswrapper[4739]: I1201 16:09:18.973875 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-84g49" podUID="ac37c1aa-2809-4754-9373-fed7588f548e" containerName="registry-server" containerID="cri-o://2b478dc2fac0dbf971ab0caff4cb683a41137faeba0cec87d8797caba652ea40" gracePeriod=2 Dec 01 16:09:19 crc kubenswrapper[4739]: I1201 16:09:19.972738 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:09:19 crc kubenswrapper[4739]: I1201 16:09:19.997345 4739 generic.go:334] "Generic (PLEG): container finished" podID="ac37c1aa-2809-4754-9373-fed7588f548e" containerID="2b478dc2fac0dbf971ab0caff4cb683a41137faeba0cec87d8797caba652ea40" exitCode=0 Dec 01 16:09:19 crc kubenswrapper[4739]: I1201 16:09:19.997403 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84g49" event={"ID":"ac37c1aa-2809-4754-9373-fed7588f548e","Type":"ContainerDied","Data":"2b478dc2fac0dbf971ab0caff4cb683a41137faeba0cec87d8797caba652ea40"} Dec 01 16:09:19 crc kubenswrapper[4739]: I1201 16:09:19.997558 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84g49" Dec 01 16:09:19 crc kubenswrapper[4739]: I1201 16:09:19.997569 4739 scope.go:117] "RemoveContainer" containerID="2b478dc2fac0dbf971ab0caff4cb683a41137faeba0cec87d8797caba652ea40" Dec 01 16:09:19 crc kubenswrapper[4739]: I1201 16:09:19.997478 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84g49" event={"ID":"ac37c1aa-2809-4754-9373-fed7588f548e","Type":"ContainerDied","Data":"f45e1e856c83b43eeadb6a9400d2234b627613312b66ba9a1f15cfde189690a8"} Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.032597 4739 scope.go:117] "RemoveContainer" containerID="c535241e2fff426d223e6b4c5ed481676d6329b6b70840614b693541b209ca41" Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.067597 4739 scope.go:117] "RemoveContainer" containerID="b03b03ed6b69cae86ec843094ee3438d62f64c36d5234512924cd1b5125a9564" Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.111860 4739 scope.go:117] "RemoveContainer" containerID="2b478dc2fac0dbf971ab0caff4cb683a41137faeba0cec87d8797caba652ea40" Dec 01 16:09:20 crc kubenswrapper[4739]: E1201 16:09:20.112502 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b478dc2fac0dbf971ab0caff4cb683a41137faeba0cec87d8797caba652ea40\": container with ID starting with 2b478dc2fac0dbf971ab0caff4cb683a41137faeba0cec87d8797caba652ea40 not found: ID does not exist" containerID="2b478dc2fac0dbf971ab0caff4cb683a41137faeba0cec87d8797caba652ea40" Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.112567 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b478dc2fac0dbf971ab0caff4cb683a41137faeba0cec87d8797caba652ea40"} err="failed to get container status \"2b478dc2fac0dbf971ab0caff4cb683a41137faeba0cec87d8797caba652ea40\": rpc error: code = NotFound desc = could not find container \"2b478dc2fac0dbf971ab0caff4cb683a41137faeba0cec87d8797caba652ea40\": container with ID starting with 2b478dc2fac0dbf971ab0caff4cb683a41137faeba0cec87d8797caba652ea40 not found: ID does not exist" Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.112609 4739 scope.go:117] "RemoveContainer" containerID="c535241e2fff426d223e6b4c5ed481676d6329b6b70840614b693541b209ca41" Dec 01 16:09:20 crc kubenswrapper[4739]: E1201 16:09:20.113173 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c535241e2fff426d223e6b4c5ed481676d6329b6b70840614b693541b209ca41\": container with ID starting with c535241e2fff426d223e6b4c5ed481676d6329b6b70840614b693541b209ca41 not found: ID does not exist" containerID="c535241e2fff426d223e6b4c5ed481676d6329b6b70840614b693541b209ca41" Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.113223 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c535241e2fff426d223e6b4c5ed481676d6329b6b70840614b693541b209ca41"} err="failed to get container status \"c535241e2fff426d223e6b4c5ed481676d6329b6b70840614b693541b209ca41\": rpc error: code = NotFound desc = could not find container \"c535241e2fff426d223e6b4c5ed481676d6329b6b70840614b693541b209ca41\": container with ID starting with c535241e2fff426d223e6b4c5ed481676d6329b6b70840614b693541b209ca41 not found: ID does not exist" Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.113249 4739 scope.go:117] "RemoveContainer" containerID="b03b03ed6b69cae86ec843094ee3438d62f64c36d5234512924cd1b5125a9564" Dec 01 16:09:20 crc kubenswrapper[4739]: E1201 16:09:20.113645 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b03b03ed6b69cae86ec843094ee3438d62f64c36d5234512924cd1b5125a9564\": container with ID starting with b03b03ed6b69cae86ec843094ee3438d62f64c36d5234512924cd1b5125a9564 not found: ID does not exist" containerID="b03b03ed6b69cae86ec843094ee3438d62f64c36d5234512924cd1b5125a9564" Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.113666 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b03b03ed6b69cae86ec843094ee3438d62f64c36d5234512924cd1b5125a9564"} err="failed to get container status \"b03b03ed6b69cae86ec843094ee3438d62f64c36d5234512924cd1b5125a9564\": rpc error: code = NotFound desc = could not find container \"b03b03ed6b69cae86ec843094ee3438d62f64c36d5234512924cd1b5125a9564\": container with ID starting with b03b03ed6b69cae86ec843094ee3438d62f64c36d5234512924cd1b5125a9564 not found: ID does not exist" Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.146755 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac37c1aa-2809-4754-9373-fed7588f548e-catalog-content\") pod \"ac37c1aa-2809-4754-9373-fed7588f548e\" (UID: \"ac37c1aa-2809-4754-9373-fed7588f548e\") " Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.147158 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac37c1aa-2809-4754-9373-fed7588f548e-utilities\") pod \"ac37c1aa-2809-4754-9373-fed7588f548e\" (UID: \"ac37c1aa-2809-4754-9373-fed7588f548e\") " Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.147616 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbntn\" (UniqueName: \"kubernetes.io/projected/ac37c1aa-2809-4754-9373-fed7588f548e-kube-api-access-xbntn\") pod \"ac37c1aa-2809-4754-9373-fed7588f548e\" (UID: \"ac37c1aa-2809-4754-9373-fed7588f548e\") " Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.148349 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac37c1aa-2809-4754-9373-fed7588f548e-utilities" (OuterVolumeSpecName: "utilities") pod "ac37c1aa-2809-4754-9373-fed7588f548e" (UID: "ac37c1aa-2809-4754-9373-fed7588f548e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.149383 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac37c1aa-2809-4754-9373-fed7588f548e-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.159648 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac37c1aa-2809-4754-9373-fed7588f548e-kube-api-access-xbntn" (OuterVolumeSpecName: "kube-api-access-xbntn") pod "ac37c1aa-2809-4754-9373-fed7588f548e" (UID: "ac37c1aa-2809-4754-9373-fed7588f548e"). InnerVolumeSpecName "kube-api-access-xbntn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.252398 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbntn\" (UniqueName: \"kubernetes.io/projected/ac37c1aa-2809-4754-9373-fed7588f548e-kube-api-access-xbntn\") on node \"crc\" DevicePath \"\"" Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.315870 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac37c1aa-2809-4754-9373-fed7588f548e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac37c1aa-2809-4754-9373-fed7588f548e" (UID: "ac37c1aa-2809-4754-9373-fed7588f548e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.353861 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac37c1aa-2809-4754-9373-fed7588f548e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.620709 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-84g49"] Dec 01 16:09:20 crc kubenswrapper[4739]: I1201 16:09:20.628032 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-84g49"] Dec 01 16:09:22 crc kubenswrapper[4739]: I1201 16:09:22.495903 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac37c1aa-2809-4754-9373-fed7588f548e" path="/var/lib/kubelet/pods/ac37c1aa-2809-4754-9373-fed7588f548e/volumes" Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.511532 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.525183 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.534156 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-7nr49"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.546178 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.554165 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.561301 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.567601 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-7nr49"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.574347 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.580874 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.587166 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fmmp5"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.593253 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bqbqx"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.599276 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zcbpd"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.605869 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z25xn"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.612177 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7jl8d"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.617679 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.624183 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.631102 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-l4rv9"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.636264 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vn45c"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.642517 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4p985"] Dec 01 16:09:39 crc kubenswrapper[4739]: I1201 16:09:39.649744 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpjn4"] Dec 01 16:09:40 crc kubenswrapper[4739]: I1201 16:09:40.491643 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b8542fe-a173-43b5-9654-89e34562f4f5" path="/var/lib/kubelet/pods/0b8542fe-a173-43b5-9654-89e34562f4f5/volumes" Dec 01 16:09:40 crc kubenswrapper[4739]: I1201 16:09:40.492483 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11851b7f-a7c3-462a-882e-a17612624b54" path="/var/lib/kubelet/pods/11851b7f-a7c3-462a-882e-a17612624b54/volumes" Dec 01 16:09:40 crc kubenswrapper[4739]: I1201 16:09:40.493210 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3027bd9d-3e05-44df-968e-228a275dcc54" path="/var/lib/kubelet/pods/3027bd9d-3e05-44df-968e-228a275dcc54/volumes" Dec 01 16:09:40 crc kubenswrapper[4739]: I1201 16:09:40.493984 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c7d3cbb-19d3-4855-8ad4-72504ba66787" path="/var/lib/kubelet/pods/5c7d3cbb-19d3-4855-8ad4-72504ba66787/volumes" Dec 01 16:09:40 crc kubenswrapper[4739]: I1201 16:09:40.495844 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ac2d686-8407-495e-b4e2-1989d8268492" path="/var/lib/kubelet/pods/6ac2d686-8407-495e-b4e2-1989d8268492/volumes" Dec 01 16:09:40 crc kubenswrapper[4739]: I1201 16:09:40.496772 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="750dd9de-c593-43c1-845f-50971d0fdb6e" path="/var/lib/kubelet/pods/750dd9de-c593-43c1-845f-50971d0fdb6e/volumes" Dec 01 16:09:40 crc kubenswrapper[4739]: I1201 16:09:40.497489 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae86132a-1c2e-4e6b-9b3a-6f8b63be4270" path="/var/lib/kubelet/pods/ae86132a-1c2e-4e6b-9b3a-6f8b63be4270/volumes" Dec 01 16:09:40 crc kubenswrapper[4739]: I1201 16:09:40.500474 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5a13eff-ceca-4a92-8cf0-8334c07e3828" path="/var/lib/kubelet/pods/b5a13eff-ceca-4a92-8cf0-8334c07e3828/volumes" Dec 01 16:09:40 crc kubenswrapper[4739]: I1201 16:09:40.501451 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed881c2f-2f6c-40d1-85dd-42c2a9393fc9" path="/var/lib/kubelet/pods/ed881c2f-2f6c-40d1-85dd-42c2a9393fc9/volumes" Dec 01 16:09:40 crc kubenswrapper[4739]: I1201 16:09:40.502449 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7b6b83c-405e-4572-ba8f-85664c4a5aba" path="/var/lib/kubelet/pods/f7b6b83c-405e-4572-ba8f-85664c4a5aba/volumes" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.295466 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n"] Dec 01 16:09:45 crc kubenswrapper[4739]: E1201 16:09:45.296524 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac37c1aa-2809-4754-9373-fed7588f548e" containerName="extract-utilities" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.296541 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac37c1aa-2809-4754-9373-fed7588f548e" containerName="extract-utilities" Dec 01 16:09:45 crc kubenswrapper[4739]: E1201 16:09:45.296564 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac37c1aa-2809-4754-9373-fed7588f548e" containerName="registry-server" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.296573 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac37c1aa-2809-4754-9373-fed7588f548e" containerName="registry-server" Dec 01 16:09:45 crc kubenswrapper[4739]: E1201 16:09:45.296600 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac37c1aa-2809-4754-9373-fed7588f548e" containerName="extract-content" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.296608 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac37c1aa-2809-4754-9373-fed7588f548e" containerName="extract-content" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.296847 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac37c1aa-2809-4754-9373-fed7588f548e" containerName="registry-server" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.297620 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.300829 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.301886 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.302298 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.302650 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.302677 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.319630 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n"] Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.367764 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.367806 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5776\" (UniqueName: \"kubernetes.io/projected/07317009-6689-44fd-8d13-1b022cc3e4d1-kube-api-access-x5776\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.367863 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.368036 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.368128 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.469777 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.469899 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.469986 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.470021 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5776\" (UniqueName: \"kubernetes.io/projected/07317009-6689-44fd-8d13-1b022cc3e4d1-kube-api-access-x5776\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.470119 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.475970 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.477628 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.480610 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.486650 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.491084 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5776\" (UniqueName: \"kubernetes.io/projected/07317009-6689-44fd-8d13-1b022cc3e4d1-kube-api-access-x5776\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:45 crc kubenswrapper[4739]: I1201 16:09:45.621908 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:09:46 crc kubenswrapper[4739]: I1201 16:09:46.216865 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n"] Dec 01 16:09:46 crc kubenswrapper[4739]: I1201 16:09:46.298923 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" event={"ID":"07317009-6689-44fd-8d13-1b022cc3e4d1","Type":"ContainerStarted","Data":"1f3f33c7dca12214fe8326b326201d6100b95c5b3c2a32ad98b45fd20fa85a11"} Dec 01 16:09:48 crc kubenswrapper[4739]: I1201 16:09:48.320187 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" event={"ID":"07317009-6689-44fd-8d13-1b022cc3e4d1","Type":"ContainerStarted","Data":"faaa64d2fe200df5ecd61a6442342737616e749aa76b83835d8c8a428e552d72"} Dec 01 16:09:48 crc kubenswrapper[4739]: I1201 16:09:48.347215 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" podStartSLOduration=2.457883507 podStartE2EDuration="3.347192253s" podCreationTimestamp="2025-12-01 16:09:45 +0000 UTC" firstStartedPulling="2025-12-01 16:09:46.226579445 +0000 UTC m=+2088.052325539" lastFinishedPulling="2025-12-01 16:09:47.115888181 +0000 UTC m=+2088.941634285" observedRunningTime="2025-12-01 16:09:48.339299629 +0000 UTC m=+2090.165045753" watchObservedRunningTime="2025-12-01 16:09:48.347192253 +0000 UTC m=+2090.172938357" Dec 01 16:10:00 crc kubenswrapper[4739]: I1201 16:10:00.434708 4739 generic.go:334] "Generic (PLEG): container finished" podID="07317009-6689-44fd-8d13-1b022cc3e4d1" containerID="faaa64d2fe200df5ecd61a6442342737616e749aa76b83835d8c8a428e552d72" exitCode=0 Dec 01 16:10:00 crc kubenswrapper[4739]: I1201 16:10:00.434827 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" event={"ID":"07317009-6689-44fd-8d13-1b022cc3e4d1","Type":"ContainerDied","Data":"faaa64d2fe200df5ecd61a6442342737616e749aa76b83835d8c8a428e552d72"} Dec 01 16:10:01 crc kubenswrapper[4739]: I1201 16:10:01.857556 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.022578 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5776\" (UniqueName: \"kubernetes.io/projected/07317009-6689-44fd-8d13-1b022cc3e4d1-kube-api-access-x5776\") pod \"07317009-6689-44fd-8d13-1b022cc3e4d1\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.022891 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-ceph\") pod \"07317009-6689-44fd-8d13-1b022cc3e4d1\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.022946 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-repo-setup-combined-ca-bundle\") pod \"07317009-6689-44fd-8d13-1b022cc3e4d1\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.022982 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-inventory\") pod \"07317009-6689-44fd-8d13-1b022cc3e4d1\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.023083 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-ssh-key\") pod \"07317009-6689-44fd-8d13-1b022cc3e4d1\" (UID: \"07317009-6689-44fd-8d13-1b022cc3e4d1\") " Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.029778 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07317009-6689-44fd-8d13-1b022cc3e4d1-kube-api-access-x5776" (OuterVolumeSpecName: "kube-api-access-x5776") pod "07317009-6689-44fd-8d13-1b022cc3e4d1" (UID: "07317009-6689-44fd-8d13-1b022cc3e4d1"). InnerVolumeSpecName "kube-api-access-x5776". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.031447 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-ceph" (OuterVolumeSpecName: "ceph") pod "07317009-6689-44fd-8d13-1b022cc3e4d1" (UID: "07317009-6689-44fd-8d13-1b022cc3e4d1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.040088 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "07317009-6689-44fd-8d13-1b022cc3e4d1" (UID: "07317009-6689-44fd-8d13-1b022cc3e4d1"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.062998 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-inventory" (OuterVolumeSpecName: "inventory") pod "07317009-6689-44fd-8d13-1b022cc3e4d1" (UID: "07317009-6689-44fd-8d13-1b022cc3e4d1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.085966 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "07317009-6689-44fd-8d13-1b022cc3e4d1" (UID: "07317009-6689-44fd-8d13-1b022cc3e4d1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.125894 4739 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.125929 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.125938 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.125946 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5776\" (UniqueName: \"kubernetes.io/projected/07317009-6689-44fd-8d13-1b022cc3e4d1-kube-api-access-x5776\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.125981 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/07317009-6689-44fd-8d13-1b022cc3e4d1-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.456528 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" event={"ID":"07317009-6689-44fd-8d13-1b022cc3e4d1","Type":"ContainerDied","Data":"1f3f33c7dca12214fe8326b326201d6100b95c5b3c2a32ad98b45fd20fa85a11"} Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.456584 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f3f33c7dca12214fe8326b326201d6100b95c5b3c2a32ad98b45fd20fa85a11" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.456596 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.558816 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w"] Dec 01 16:10:02 crc kubenswrapper[4739]: E1201 16:10:02.559244 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07317009-6689-44fd-8d13-1b022cc3e4d1" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.559263 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="07317009-6689-44fd-8d13-1b022cc3e4d1" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.559527 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="07317009-6689-44fd-8d13-1b022cc3e4d1" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.560241 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.564600 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.564959 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.565175 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.565363 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.572319 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.573319 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w"] Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.735731 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klsrd\" (UniqueName: \"kubernetes.io/projected/df49ac71-260c-49a1-ac48-ab9e863e08be-kube-api-access-klsrd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.736147 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.736305 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.736514 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.736695 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.838405 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.838520 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.838569 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klsrd\" (UniqueName: \"kubernetes.io/projected/df49ac71-260c-49a1-ac48-ab9e863e08be-kube-api-access-klsrd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.838638 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.838684 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.844808 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.846563 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.846963 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.849922 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.864870 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klsrd\" (UniqueName: \"kubernetes.io/projected/df49ac71-260c-49a1-ac48-ab9e863e08be-kube-api-access-klsrd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:02 crc kubenswrapper[4739]: I1201 16:10:02.883953 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:10:03 crc kubenswrapper[4739]: I1201 16:10:03.565120 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w"] Dec 01 16:10:03 crc kubenswrapper[4739]: W1201 16:10:03.569016 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf49ac71_260c_49a1_ac48_ab9e863e08be.slice/crio-9fdc3338d032617c06f9f9b4800482c9023036ec5bf9f580468342f6d897776d WatchSource:0}: Error finding container 9fdc3338d032617c06f9f9b4800482c9023036ec5bf9f580468342f6d897776d: Status 404 returned error can't find the container with id 9fdc3338d032617c06f9f9b4800482c9023036ec5bf9f580468342f6d897776d Dec 01 16:10:04 crc kubenswrapper[4739]: I1201 16:10:04.488797 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" event={"ID":"df49ac71-260c-49a1-ac48-ab9e863e08be","Type":"ContainerStarted","Data":"9fdc3338d032617c06f9f9b4800482c9023036ec5bf9f580468342f6d897776d"} Dec 01 16:10:05 crc kubenswrapper[4739]: I1201 16:10:05.492321 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" event={"ID":"df49ac71-260c-49a1-ac48-ab9e863e08be","Type":"ContainerStarted","Data":"b03706e12f26ecf4fc22a4154642e4508dd119fdd4aa845e77898734bfef883b"} Dec 01 16:10:05 crc kubenswrapper[4739]: I1201 16:10:05.517883 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" podStartSLOduration=2.77905318 podStartE2EDuration="3.517857226s" podCreationTimestamp="2025-12-01 16:10:02 +0000 UTC" firstStartedPulling="2025-12-01 16:10:03.57147834 +0000 UTC m=+2105.397224434" lastFinishedPulling="2025-12-01 16:10:04.310282386 +0000 UTC m=+2106.136028480" observedRunningTime="2025-12-01 16:10:05.509282031 +0000 UTC m=+2107.335028165" watchObservedRunningTime="2025-12-01 16:10:05.517857226 +0000 UTC m=+2107.343603330" Dec 01 16:10:09 crc kubenswrapper[4739]: I1201 16:10:09.621595 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:10:09 crc kubenswrapper[4739]: I1201 16:10:09.622082 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:10:22 crc kubenswrapper[4739]: I1201 16:10:22.893839 4739 scope.go:117] "RemoveContainer" containerID="1ce022649232a707c748a1a9afd1480938b067c019eec773ab1b09bdc08f8a1c" Dec 01 16:10:22 crc kubenswrapper[4739]: I1201 16:10:22.949902 4739 scope.go:117] "RemoveContainer" containerID="01c05052ed43ba4db45fb5369fabc3240bf79c40f3bbe78d9d3dd947bc318eb3" Dec 01 16:10:22 crc kubenswrapper[4739]: I1201 16:10:22.996878 4739 scope.go:117] "RemoveContainer" containerID="2d1187e9457eac46edf1a1c390e3a91f1575ec4d78e403ede67f59bc9d257eb6" Dec 01 16:10:23 crc kubenswrapper[4739]: I1201 16:10:23.053131 4739 scope.go:117] "RemoveContainer" containerID="7b6e73dee68461b7d90ae8d167ee91b2fed52c706164e5fafbe31e652dbb61c5" Dec 01 16:10:23 crc kubenswrapper[4739]: I1201 16:10:23.097691 4739 scope.go:117] "RemoveContainer" containerID="577c84f6a4649de3d4b7a9eeca3414fbf7ae3bb3c7fe99e1e794d138ad80aab6" Dec 01 16:10:23 crc kubenswrapper[4739]: I1201 16:10:23.127930 4739 scope.go:117] "RemoveContainer" containerID="cb26e7e043112190326b4e04beaaab20e5f5d7b71e3274844c6b90e1c92d063b" Dec 01 16:10:23 crc kubenswrapper[4739]: I1201 16:10:23.217679 4739 scope.go:117] "RemoveContainer" containerID="3fac8a90bed2abfaa40fef7599e3db6a3a00c3706112ac25c5890105b83ed039" Dec 01 16:10:23 crc kubenswrapper[4739]: I1201 16:10:23.262632 4739 scope.go:117] "RemoveContainer" containerID="bf5451c2fa5291fc98935f3660e6f7c55c1e9340b98cb227a0f4a27326da5e30" Dec 01 16:10:23 crc kubenswrapper[4739]: I1201 16:10:23.287081 4739 scope.go:117] "RemoveContainer" containerID="f804a154b9c41c7489a2b8af6e2aa383bf99e78ecd75cb2c7f48ee8c24c89faf" Dec 01 16:10:39 crc kubenswrapper[4739]: I1201 16:10:39.621915 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:10:39 crc kubenswrapper[4739]: I1201 16:10:39.622611 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:10:47 crc kubenswrapper[4739]: I1201 16:10:47.789647 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zr2cw"] Dec 01 16:10:47 crc kubenswrapper[4739]: I1201 16:10:47.791707 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:10:47 crc kubenswrapper[4739]: I1201 16:10:47.811757 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zr2cw"] Dec 01 16:10:47 crc kubenswrapper[4739]: I1201 16:10:47.957273 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vts4\" (UniqueName: \"kubernetes.io/projected/ac1af94a-242a-4843-a38f-18c79ff47545-kube-api-access-7vts4\") pod \"community-operators-zr2cw\" (UID: \"ac1af94a-242a-4843-a38f-18c79ff47545\") " pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:10:47 crc kubenswrapper[4739]: I1201 16:10:47.957436 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac1af94a-242a-4843-a38f-18c79ff47545-utilities\") pod \"community-operators-zr2cw\" (UID: \"ac1af94a-242a-4843-a38f-18c79ff47545\") " pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:10:47 crc kubenswrapper[4739]: I1201 16:10:47.957484 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac1af94a-242a-4843-a38f-18c79ff47545-catalog-content\") pod \"community-operators-zr2cw\" (UID: \"ac1af94a-242a-4843-a38f-18c79ff47545\") " pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:10:48 crc kubenswrapper[4739]: I1201 16:10:48.059276 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac1af94a-242a-4843-a38f-18c79ff47545-utilities\") pod \"community-operators-zr2cw\" (UID: \"ac1af94a-242a-4843-a38f-18c79ff47545\") " pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:10:48 crc kubenswrapper[4739]: I1201 16:10:48.059375 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac1af94a-242a-4843-a38f-18c79ff47545-catalog-content\") pod \"community-operators-zr2cw\" (UID: \"ac1af94a-242a-4843-a38f-18c79ff47545\") " pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:10:48 crc kubenswrapper[4739]: I1201 16:10:48.059488 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vts4\" (UniqueName: \"kubernetes.io/projected/ac1af94a-242a-4843-a38f-18c79ff47545-kube-api-access-7vts4\") pod \"community-operators-zr2cw\" (UID: \"ac1af94a-242a-4843-a38f-18c79ff47545\") " pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:10:48 crc kubenswrapper[4739]: I1201 16:10:48.059981 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac1af94a-242a-4843-a38f-18c79ff47545-catalog-content\") pod \"community-operators-zr2cw\" (UID: \"ac1af94a-242a-4843-a38f-18c79ff47545\") " pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:10:48 crc kubenswrapper[4739]: I1201 16:10:48.060383 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac1af94a-242a-4843-a38f-18c79ff47545-utilities\") pod \"community-operators-zr2cw\" (UID: \"ac1af94a-242a-4843-a38f-18c79ff47545\") " pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:10:48 crc kubenswrapper[4739]: I1201 16:10:48.086544 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vts4\" (UniqueName: \"kubernetes.io/projected/ac1af94a-242a-4843-a38f-18c79ff47545-kube-api-access-7vts4\") pod \"community-operators-zr2cw\" (UID: \"ac1af94a-242a-4843-a38f-18c79ff47545\") " pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:10:48 crc kubenswrapper[4739]: I1201 16:10:48.122334 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:10:48 crc kubenswrapper[4739]: I1201 16:10:48.597983 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zr2cw"] Dec 01 16:10:48 crc kubenswrapper[4739]: I1201 16:10:48.896405 4739 generic.go:334] "Generic (PLEG): container finished" podID="ac1af94a-242a-4843-a38f-18c79ff47545" containerID="d964b9430882d08a2713543c5a90d0306bd4bc264fb08a60b29f7b80dbcf139c" exitCode=0 Dec 01 16:10:48 crc kubenswrapper[4739]: I1201 16:10:48.896693 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zr2cw" event={"ID":"ac1af94a-242a-4843-a38f-18c79ff47545","Type":"ContainerDied","Data":"d964b9430882d08a2713543c5a90d0306bd4bc264fb08a60b29f7b80dbcf139c"} Dec 01 16:10:48 crc kubenswrapper[4739]: I1201 16:10:48.896717 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zr2cw" event={"ID":"ac1af94a-242a-4843-a38f-18c79ff47545","Type":"ContainerStarted","Data":"5e85af60a1092b228c362ff45087f1efef1b4a79ce7d44b95846aff4b5b4f4ac"} Dec 01 16:10:50 crc kubenswrapper[4739]: I1201 16:10:50.920253 4739 generic.go:334] "Generic (PLEG): container finished" podID="ac1af94a-242a-4843-a38f-18c79ff47545" containerID="c66447720c67518718bef36ffb29039f9a6b2dad3fbd8f5010a68f8761921e7b" exitCode=0 Dec 01 16:10:50 crc kubenswrapper[4739]: I1201 16:10:50.920375 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zr2cw" event={"ID":"ac1af94a-242a-4843-a38f-18c79ff47545","Type":"ContainerDied","Data":"c66447720c67518718bef36ffb29039f9a6b2dad3fbd8f5010a68f8761921e7b"} Dec 01 16:10:53 crc kubenswrapper[4739]: I1201 16:10:53.145665 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zr2cw" event={"ID":"ac1af94a-242a-4843-a38f-18c79ff47545","Type":"ContainerStarted","Data":"da0ff1e24609ba85ac1dec75ad63dd98c6902b03b31d485130ea4f41fb1f0942"} Dec 01 16:10:53 crc kubenswrapper[4739]: I1201 16:10:53.181274 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zr2cw" podStartSLOduration=2.902914543 podStartE2EDuration="6.18125285s" podCreationTimestamp="2025-12-01 16:10:47 +0000 UTC" firstStartedPulling="2025-12-01 16:10:48.898524511 +0000 UTC m=+2150.724270605" lastFinishedPulling="2025-12-01 16:10:52.176862808 +0000 UTC m=+2154.002608912" observedRunningTime="2025-12-01 16:10:53.169763015 +0000 UTC m=+2154.995509129" watchObservedRunningTime="2025-12-01 16:10:53.18125285 +0000 UTC m=+2155.006998954" Dec 01 16:10:58 crc kubenswrapper[4739]: I1201 16:10:58.126044 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:10:58 crc kubenswrapper[4739]: I1201 16:10:58.126543 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:10:58 crc kubenswrapper[4739]: I1201 16:10:58.182168 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:10:58 crc kubenswrapper[4739]: I1201 16:10:58.241520 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:10:58 crc kubenswrapper[4739]: I1201 16:10:58.421484 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zr2cw"] Dec 01 16:11:00 crc kubenswrapper[4739]: I1201 16:11:00.216288 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zr2cw" podUID="ac1af94a-242a-4843-a38f-18c79ff47545" containerName="registry-server" containerID="cri-o://da0ff1e24609ba85ac1dec75ad63dd98c6902b03b31d485130ea4f41fb1f0942" gracePeriod=2 Dec 01 16:11:01 crc kubenswrapper[4739]: I1201 16:11:01.235923 4739 generic.go:334] "Generic (PLEG): container finished" podID="ac1af94a-242a-4843-a38f-18c79ff47545" containerID="da0ff1e24609ba85ac1dec75ad63dd98c6902b03b31d485130ea4f41fb1f0942" exitCode=0 Dec 01 16:11:01 crc kubenswrapper[4739]: I1201 16:11:01.235979 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zr2cw" event={"ID":"ac1af94a-242a-4843-a38f-18c79ff47545","Type":"ContainerDied","Data":"da0ff1e24609ba85ac1dec75ad63dd98c6902b03b31d485130ea4f41fb1f0942"} Dec 01 16:11:01 crc kubenswrapper[4739]: I1201 16:11:01.413663 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:11:01 crc kubenswrapper[4739]: I1201 16:11:01.557554 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vts4\" (UniqueName: \"kubernetes.io/projected/ac1af94a-242a-4843-a38f-18c79ff47545-kube-api-access-7vts4\") pod \"ac1af94a-242a-4843-a38f-18c79ff47545\" (UID: \"ac1af94a-242a-4843-a38f-18c79ff47545\") " Dec 01 16:11:01 crc kubenswrapper[4739]: I1201 16:11:01.557830 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac1af94a-242a-4843-a38f-18c79ff47545-utilities\") pod \"ac1af94a-242a-4843-a38f-18c79ff47545\" (UID: \"ac1af94a-242a-4843-a38f-18c79ff47545\") " Dec 01 16:11:01 crc kubenswrapper[4739]: I1201 16:11:01.557917 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac1af94a-242a-4843-a38f-18c79ff47545-catalog-content\") pod \"ac1af94a-242a-4843-a38f-18c79ff47545\" (UID: \"ac1af94a-242a-4843-a38f-18c79ff47545\") " Dec 01 16:11:01 crc kubenswrapper[4739]: I1201 16:11:01.559122 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac1af94a-242a-4843-a38f-18c79ff47545-utilities" (OuterVolumeSpecName: "utilities") pod "ac1af94a-242a-4843-a38f-18c79ff47545" (UID: "ac1af94a-242a-4843-a38f-18c79ff47545"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:11:01 crc kubenswrapper[4739]: I1201 16:11:01.567624 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac1af94a-242a-4843-a38f-18c79ff47545-kube-api-access-7vts4" (OuterVolumeSpecName: "kube-api-access-7vts4") pod "ac1af94a-242a-4843-a38f-18c79ff47545" (UID: "ac1af94a-242a-4843-a38f-18c79ff47545"). InnerVolumeSpecName "kube-api-access-7vts4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:11:01 crc kubenswrapper[4739]: I1201 16:11:01.618146 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac1af94a-242a-4843-a38f-18c79ff47545-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac1af94a-242a-4843-a38f-18c79ff47545" (UID: "ac1af94a-242a-4843-a38f-18c79ff47545"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:11:01 crc kubenswrapper[4739]: I1201 16:11:01.660668 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac1af94a-242a-4843-a38f-18c79ff47545-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:01 crc kubenswrapper[4739]: I1201 16:11:01.660717 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac1af94a-242a-4843-a38f-18c79ff47545-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:01 crc kubenswrapper[4739]: I1201 16:11:01.660755 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vts4\" (UniqueName: \"kubernetes.io/projected/ac1af94a-242a-4843-a38f-18c79ff47545-kube-api-access-7vts4\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:02 crc kubenswrapper[4739]: I1201 16:11:02.248927 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zr2cw" event={"ID":"ac1af94a-242a-4843-a38f-18c79ff47545","Type":"ContainerDied","Data":"5e85af60a1092b228c362ff45087f1efef1b4a79ce7d44b95846aff4b5b4f4ac"} Dec 01 16:11:02 crc kubenswrapper[4739]: I1201 16:11:02.248990 4739 scope.go:117] "RemoveContainer" containerID="da0ff1e24609ba85ac1dec75ad63dd98c6902b03b31d485130ea4f41fb1f0942" Dec 01 16:11:02 crc kubenswrapper[4739]: I1201 16:11:02.249007 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zr2cw" Dec 01 16:11:02 crc kubenswrapper[4739]: I1201 16:11:02.275263 4739 scope.go:117] "RemoveContainer" containerID="c66447720c67518718bef36ffb29039f9a6b2dad3fbd8f5010a68f8761921e7b" Dec 01 16:11:02 crc kubenswrapper[4739]: I1201 16:11:02.294212 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zr2cw"] Dec 01 16:11:02 crc kubenswrapper[4739]: I1201 16:11:02.301468 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zr2cw"] Dec 01 16:11:02 crc kubenswrapper[4739]: I1201 16:11:02.314897 4739 scope.go:117] "RemoveContainer" containerID="d964b9430882d08a2713543c5a90d0306bd4bc264fb08a60b29f7b80dbcf139c" Dec 01 16:11:02 crc kubenswrapper[4739]: I1201 16:11:02.494711 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac1af94a-242a-4843-a38f-18c79ff47545" path="/var/lib/kubelet/pods/ac1af94a-242a-4843-a38f-18c79ff47545/volumes" Dec 01 16:11:09 crc kubenswrapper[4739]: I1201 16:11:09.622388 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:11:09 crc kubenswrapper[4739]: I1201 16:11:09.623186 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:11:09 crc kubenswrapper[4739]: I1201 16:11:09.623254 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 16:11:09 crc kubenswrapper[4739]: I1201 16:11:09.624196 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:11:09 crc kubenswrapper[4739]: I1201 16:11:09.624278 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" gracePeriod=600 Dec 01 16:11:09 crc kubenswrapper[4739]: E1201 16:11:09.755131 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:11:10 crc kubenswrapper[4739]: I1201 16:11:10.352059 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" exitCode=0 Dec 01 16:11:10 crc kubenswrapper[4739]: I1201 16:11:10.352116 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be"} Dec 01 16:11:10 crc kubenswrapper[4739]: I1201 16:11:10.352259 4739 scope.go:117] "RemoveContainer" containerID="ddde4c9dc4ce14363e987196bc7779dbb2cfefa0de2f710f51b5795ef1b081bd" Dec 01 16:11:10 crc kubenswrapper[4739]: I1201 16:11:10.353191 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:11:10 crc kubenswrapper[4739]: E1201 16:11:10.353807 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:11:21 crc kubenswrapper[4739]: I1201 16:11:21.476819 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:11:21 crc kubenswrapper[4739]: E1201 16:11:21.478654 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:11:23 crc kubenswrapper[4739]: I1201 16:11:23.560610 4739 scope.go:117] "RemoveContainer" containerID="b3bb33d102750bb8da2d403c3a0abe849107eede5f0b0b87d7010ea788e82eb2" Dec 01 16:11:34 crc kubenswrapper[4739]: I1201 16:11:34.478410 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:11:34 crc kubenswrapper[4739]: E1201 16:11:34.481779 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:11:45 crc kubenswrapper[4739]: I1201 16:11:45.477382 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:11:45 crc kubenswrapper[4739]: E1201 16:11:45.478067 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:11:46 crc kubenswrapper[4739]: I1201 16:11:46.701537 4739 generic.go:334] "Generic (PLEG): container finished" podID="df49ac71-260c-49a1-ac48-ab9e863e08be" containerID="b03706e12f26ecf4fc22a4154642e4508dd119fdd4aa845e77898734bfef883b" exitCode=0 Dec 01 16:11:46 crc kubenswrapper[4739]: I1201 16:11:46.701684 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" event={"ID":"df49ac71-260c-49a1-ac48-ab9e863e08be","Type":"ContainerDied","Data":"b03706e12f26ecf4fc22a4154642e4508dd119fdd4aa845e77898734bfef883b"} Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.173937 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.263871 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-ceph\") pod \"df49ac71-260c-49a1-ac48-ab9e863e08be\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.264409 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-inventory\") pod \"df49ac71-260c-49a1-ac48-ab9e863e08be\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.264655 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klsrd\" (UniqueName: \"kubernetes.io/projected/df49ac71-260c-49a1-ac48-ab9e863e08be-kube-api-access-klsrd\") pod \"df49ac71-260c-49a1-ac48-ab9e863e08be\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.264808 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-bootstrap-combined-ca-bundle\") pod \"df49ac71-260c-49a1-ac48-ab9e863e08be\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.265017 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-ssh-key\") pod \"df49ac71-260c-49a1-ac48-ab9e863e08be\" (UID: \"df49ac71-260c-49a1-ac48-ab9e863e08be\") " Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.270320 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-ceph" (OuterVolumeSpecName: "ceph") pod "df49ac71-260c-49a1-ac48-ab9e863e08be" (UID: "df49ac71-260c-49a1-ac48-ab9e863e08be"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.270411 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df49ac71-260c-49a1-ac48-ab9e863e08be-kube-api-access-klsrd" (OuterVolumeSpecName: "kube-api-access-klsrd") pod "df49ac71-260c-49a1-ac48-ab9e863e08be" (UID: "df49ac71-260c-49a1-ac48-ab9e863e08be"). InnerVolumeSpecName "kube-api-access-klsrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.271141 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "df49ac71-260c-49a1-ac48-ab9e863e08be" (UID: "df49ac71-260c-49a1-ac48-ab9e863e08be"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.295385 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "df49ac71-260c-49a1-ac48-ab9e863e08be" (UID: "df49ac71-260c-49a1-ac48-ab9e863e08be"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.306993 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-inventory" (OuterVolumeSpecName: "inventory") pod "df49ac71-260c-49a1-ac48-ab9e863e08be" (UID: "df49ac71-260c-49a1-ac48-ab9e863e08be"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.367972 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.368018 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klsrd\" (UniqueName: \"kubernetes.io/projected/df49ac71-260c-49a1-ac48-ab9e863e08be-kube-api-access-klsrd\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.368034 4739 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.368048 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.368061 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df49ac71-260c-49a1-ac48-ab9e863e08be-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.672847 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-48dnc"] Dec 01 16:11:48 crc kubenswrapper[4739]: E1201 16:11:48.673584 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac1af94a-242a-4843-a38f-18c79ff47545" containerName="registry-server" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.673603 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac1af94a-242a-4843-a38f-18c79ff47545" containerName="registry-server" Dec 01 16:11:48 crc kubenswrapper[4739]: E1201 16:11:48.673617 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac1af94a-242a-4843-a38f-18c79ff47545" containerName="extract-utilities" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.673625 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac1af94a-242a-4843-a38f-18c79ff47545" containerName="extract-utilities" Dec 01 16:11:48 crc kubenswrapper[4739]: E1201 16:11:48.673662 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df49ac71-260c-49a1-ac48-ab9e863e08be" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.673672 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="df49ac71-260c-49a1-ac48-ab9e863e08be" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 16:11:48 crc kubenswrapper[4739]: E1201 16:11:48.673689 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac1af94a-242a-4843-a38f-18c79ff47545" containerName="extract-content" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.673696 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac1af94a-242a-4843-a38f-18c79ff47545" containerName="extract-content" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.673909 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="df49ac71-260c-49a1-ac48-ab9e863e08be" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.673948 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac1af94a-242a-4843-a38f-18c79ff47545" containerName="registry-server" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.675522 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.713524 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-48dnc"] Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.724033 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" event={"ID":"df49ac71-260c-49a1-ac48-ab9e863e08be","Type":"ContainerDied","Data":"9fdc3338d032617c06f9f9b4800482c9023036ec5bf9f580468342f6d897776d"} Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.724087 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9fdc3338d032617c06f9f9b4800482c9023036ec5bf9f580468342f6d897776d" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.724155 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.776361 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-utilities\") pod \"redhat-marketplace-48dnc\" (UID: \"97c5fbe3-8871-4cbc-885a-d02783e3fc4c\") " pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.776698 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhl9n\" (UniqueName: \"kubernetes.io/projected/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-kube-api-access-fhl9n\") pod \"redhat-marketplace-48dnc\" (UID: \"97c5fbe3-8871-4cbc-885a-d02783e3fc4c\") " pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.776821 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-catalog-content\") pod \"redhat-marketplace-48dnc\" (UID: \"97c5fbe3-8871-4cbc-885a-d02783e3fc4c\") " pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.808198 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m"] Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.809745 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.812591 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.813020 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.813243 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.813432 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.814702 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.820392 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m"] Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.878546 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhl9n\" (UniqueName: \"kubernetes.io/projected/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-kube-api-access-fhl9n\") pod \"redhat-marketplace-48dnc\" (UID: \"97c5fbe3-8871-4cbc-885a-d02783e3fc4c\") " pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.878611 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-catalog-content\") pod \"redhat-marketplace-48dnc\" (UID: \"97c5fbe3-8871-4cbc-885a-d02783e3fc4c\") " pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.878671 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-utilities\") pod \"redhat-marketplace-48dnc\" (UID: \"97c5fbe3-8871-4cbc-885a-d02783e3fc4c\") " pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.879080 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-utilities\") pod \"redhat-marketplace-48dnc\" (UID: \"97c5fbe3-8871-4cbc-885a-d02783e3fc4c\") " pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.879245 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-catalog-content\") pod \"redhat-marketplace-48dnc\" (UID: \"97c5fbe3-8871-4cbc-885a-d02783e3fc4c\") " pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.896471 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhl9n\" (UniqueName: \"kubernetes.io/projected/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-kube-api-access-fhl9n\") pod \"redhat-marketplace-48dnc\" (UID: \"97c5fbe3-8871-4cbc-885a-d02783e3fc4c\") " pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.980453 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.980517 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxs7b\" (UniqueName: \"kubernetes.io/projected/0c99055a-3919-465d-8584-a82a814a4be7-kube-api-access-jxs7b\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.980573 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:11:48 crc kubenswrapper[4739]: I1201 16:11:48.980604 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.005651 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.083169 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.083556 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxs7b\" (UniqueName: \"kubernetes.io/projected/0c99055a-3919-465d-8584-a82a814a4be7-kube-api-access-jxs7b\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.083603 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.083635 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.092719 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.093127 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.119238 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.178360 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxs7b\" (UniqueName: \"kubernetes.io/projected/0c99055a-3919-465d-8584-a82a814a4be7-kube-api-access-jxs7b\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.249506 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vsxbw"] Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.259494 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.264645 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vsxbw"] Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.388964 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-utilities\") pod \"certified-operators-vsxbw\" (UID: \"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02\") " pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.389062 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md9j4\" (UniqueName: \"kubernetes.io/projected/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-kube-api-access-md9j4\") pod \"certified-operators-vsxbw\" (UID: \"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02\") " pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.389125 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-catalog-content\") pod \"certified-operators-vsxbw\" (UID: \"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02\") " pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.427686 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.490520 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-utilities\") pod \"certified-operators-vsxbw\" (UID: \"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02\") " pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.491449 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-utilities\") pod \"certified-operators-vsxbw\" (UID: \"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02\") " pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.491604 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md9j4\" (UniqueName: \"kubernetes.io/projected/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-kube-api-access-md9j4\") pod \"certified-operators-vsxbw\" (UID: \"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02\") " pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.492676 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-catalog-content\") pod \"certified-operators-vsxbw\" (UID: \"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02\") " pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.493143 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-catalog-content\") pod \"certified-operators-vsxbw\" (UID: \"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02\") " pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.520712 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md9j4\" (UniqueName: \"kubernetes.io/projected/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-kube-api-access-md9j4\") pod \"certified-operators-vsxbw\" (UID: \"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02\") " pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.558309 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-48dnc"] Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.600632 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.761156 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-48dnc" event={"ID":"97c5fbe3-8871-4cbc-885a-d02783e3fc4c","Type":"ContainerStarted","Data":"1ac069673522aefc254e07eea89578d34f214fe02f11438def83edc87af75de8"} Dec 01 16:11:49 crc kubenswrapper[4739]: I1201 16:11:49.993347 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m"] Dec 01 16:11:50 crc kubenswrapper[4739]: W1201 16:11:50.131051 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d73bd1d_f9f6_4bef_8fe7_06f6a67b1a02.slice/crio-5edb755d47ba8408734633a728a4b1083973c0db6644884852e0aea796a7530f WatchSource:0}: Error finding container 5edb755d47ba8408734633a728a4b1083973c0db6644884852e0aea796a7530f: Status 404 returned error can't find the container with id 5edb755d47ba8408734633a728a4b1083973c0db6644884852e0aea796a7530f Dec 01 16:11:50 crc kubenswrapper[4739]: I1201 16:11:50.131169 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vsxbw"] Dec 01 16:11:50 crc kubenswrapper[4739]: I1201 16:11:50.775088 4739 generic.go:334] "Generic (PLEG): container finished" podID="6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02" containerID="a3b9db03ccba1e5150a66bd57116db3b2554e5bab18d25d56486e216dd26f250" exitCode=0 Dec 01 16:11:50 crc kubenswrapper[4739]: I1201 16:11:50.775179 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsxbw" event={"ID":"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02","Type":"ContainerDied","Data":"a3b9db03ccba1e5150a66bd57116db3b2554e5bab18d25d56486e216dd26f250"} Dec 01 16:11:50 crc kubenswrapper[4739]: I1201 16:11:50.775510 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsxbw" event={"ID":"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02","Type":"ContainerStarted","Data":"5edb755d47ba8408734633a728a4b1083973c0db6644884852e0aea796a7530f"} Dec 01 16:11:50 crc kubenswrapper[4739]: I1201 16:11:50.778931 4739 generic.go:334] "Generic (PLEG): container finished" podID="97c5fbe3-8871-4cbc-885a-d02783e3fc4c" containerID="d75917330552ef8ac38f1bd2ebba3afbd7c0f8a8eb031459d4e841cf382c66ac" exitCode=0 Dec 01 16:11:50 crc kubenswrapper[4739]: I1201 16:11:50.779066 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-48dnc" event={"ID":"97c5fbe3-8871-4cbc-885a-d02783e3fc4c","Type":"ContainerDied","Data":"d75917330552ef8ac38f1bd2ebba3afbd7c0f8a8eb031459d4e841cf382c66ac"} Dec 01 16:11:50 crc kubenswrapper[4739]: I1201 16:11:50.782116 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" event={"ID":"0c99055a-3919-465d-8584-a82a814a4be7","Type":"ContainerStarted","Data":"d7c00ed0a39f8179b9b0cc533f093e3ae81060fe4c27f0dd785b3ff54a49bc1f"} Dec 01 16:11:51 crc kubenswrapper[4739]: I1201 16:11:51.790229 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" event={"ID":"0c99055a-3919-465d-8584-a82a814a4be7","Type":"ContainerStarted","Data":"142a149e28f4080bf1114bdd56960de36aeb1a277453f032847583c620ed55c1"} Dec 01 16:11:51 crc kubenswrapper[4739]: I1201 16:11:51.795859 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-48dnc" event={"ID":"97c5fbe3-8871-4cbc-885a-d02783e3fc4c","Type":"ContainerStarted","Data":"4924ec711b09779cc99715dfef504021763fcec7440053380ca09676c71078ea"} Dec 01 16:11:51 crc kubenswrapper[4739]: I1201 16:11:51.819594 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" podStartSLOduration=3.144276316 podStartE2EDuration="3.819564939s" podCreationTimestamp="2025-12-01 16:11:48 +0000 UTC" firstStartedPulling="2025-12-01 16:11:49.992971585 +0000 UTC m=+2211.818717679" lastFinishedPulling="2025-12-01 16:11:50.668260168 +0000 UTC m=+2212.494006302" observedRunningTime="2025-12-01 16:11:51.807551558 +0000 UTC m=+2213.633297692" watchObservedRunningTime="2025-12-01 16:11:51.819564939 +0000 UTC m=+2213.645311063" Dec 01 16:11:52 crc kubenswrapper[4739]: I1201 16:11:52.810799 4739 generic.go:334] "Generic (PLEG): container finished" podID="97c5fbe3-8871-4cbc-885a-d02783e3fc4c" containerID="4924ec711b09779cc99715dfef504021763fcec7440053380ca09676c71078ea" exitCode=0 Dec 01 16:11:52 crc kubenswrapper[4739]: I1201 16:11:52.810913 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-48dnc" event={"ID":"97c5fbe3-8871-4cbc-885a-d02783e3fc4c","Type":"ContainerDied","Data":"4924ec711b09779cc99715dfef504021763fcec7440053380ca09676c71078ea"} Dec 01 16:11:52 crc kubenswrapper[4739]: I1201 16:11:52.814384 4739 generic.go:334] "Generic (PLEG): container finished" podID="6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02" containerID="42947a3ac64c5ad508fa2e43a0b6f7f2ba1d3d66dfdd477f2eb9a5cf432071f0" exitCode=0 Dec 01 16:11:52 crc kubenswrapper[4739]: I1201 16:11:52.814497 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsxbw" event={"ID":"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02","Type":"ContainerDied","Data":"42947a3ac64c5ad508fa2e43a0b6f7f2ba1d3d66dfdd477f2eb9a5cf432071f0"} Dec 01 16:11:53 crc kubenswrapper[4739]: I1201 16:11:53.835970 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsxbw" event={"ID":"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02","Type":"ContainerStarted","Data":"4b3ae247a77b9ca86a18b68baa10e2405c6aa764dbf08e992ffd9b53ab46525d"} Dec 01 16:11:53 crc kubenswrapper[4739]: I1201 16:11:53.872169 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vsxbw" podStartSLOduration=2.313664069 podStartE2EDuration="4.872148915s" podCreationTimestamp="2025-12-01 16:11:49 +0000 UTC" firstStartedPulling="2025-12-01 16:11:50.778579847 +0000 UTC m=+2212.604325951" lastFinishedPulling="2025-12-01 16:11:53.337064673 +0000 UTC m=+2215.162810797" observedRunningTime="2025-12-01 16:11:53.859802084 +0000 UTC m=+2215.685548178" watchObservedRunningTime="2025-12-01 16:11:53.872148915 +0000 UTC m=+2215.697895009" Dec 01 16:11:54 crc kubenswrapper[4739]: I1201 16:11:54.848038 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-48dnc" event={"ID":"97c5fbe3-8871-4cbc-885a-d02783e3fc4c","Type":"ContainerStarted","Data":"911a305809350d8657e5588469d216b7928791dac03e33a4f8d08490c0136ae4"} Dec 01 16:11:54 crc kubenswrapper[4739]: I1201 16:11:54.875563 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-48dnc" podStartSLOduration=4.019557597 podStartE2EDuration="6.875546616s" podCreationTimestamp="2025-12-01 16:11:48 +0000 UTC" firstStartedPulling="2025-12-01 16:11:50.780834106 +0000 UTC m=+2212.606580210" lastFinishedPulling="2025-12-01 16:11:53.636823095 +0000 UTC m=+2215.462569229" observedRunningTime="2025-12-01 16:11:54.865349251 +0000 UTC m=+2216.691095355" watchObservedRunningTime="2025-12-01 16:11:54.875546616 +0000 UTC m=+2216.701292710" Dec 01 16:11:58 crc kubenswrapper[4739]: I1201 16:11:58.488612 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:11:58 crc kubenswrapper[4739]: E1201 16:11:58.491730 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:11:59 crc kubenswrapper[4739]: I1201 16:11:59.006800 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:11:59 crc kubenswrapper[4739]: I1201 16:11:59.007115 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:11:59 crc kubenswrapper[4739]: I1201 16:11:59.059154 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:11:59 crc kubenswrapper[4739]: I1201 16:11:59.601444 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:11:59 crc kubenswrapper[4739]: I1201 16:11:59.601754 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:11:59 crc kubenswrapper[4739]: I1201 16:11:59.650316 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:11:59 crc kubenswrapper[4739]: I1201 16:11:59.956959 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:11:59 crc kubenswrapper[4739]: I1201 16:11:59.964067 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:12:00 crc kubenswrapper[4739]: I1201 16:12:00.439072 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vsxbw"] Dec 01 16:12:01 crc kubenswrapper[4739]: I1201 16:12:01.918140 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vsxbw" podUID="6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02" containerName="registry-server" containerID="cri-o://4b3ae247a77b9ca86a18b68baa10e2405c6aa764dbf08e992ffd9b53ab46525d" gracePeriod=2 Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.237067 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-48dnc"] Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.237622 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-48dnc" podUID="97c5fbe3-8871-4cbc-885a-d02783e3fc4c" containerName="registry-server" containerID="cri-o://911a305809350d8657e5588469d216b7928791dac03e33a4f8d08490c0136ae4" gracePeriod=2 Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.879585 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.919335 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-catalog-content\") pod \"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02\" (UID: \"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02\") " Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.919387 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md9j4\" (UniqueName: \"kubernetes.io/projected/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-kube-api-access-md9j4\") pod \"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02\" (UID: \"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02\") " Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.919490 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-utilities\") pod \"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02\" (UID: \"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02\") " Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.923625 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-utilities" (OuterVolumeSpecName: "utilities") pod "6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02" (UID: "6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.938531 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-kube-api-access-md9j4" (OuterVolumeSpecName: "kube-api-access-md9j4") pod "6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02" (UID: "6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02"). InnerVolumeSpecName "kube-api-access-md9j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.940072 4739 generic.go:334] "Generic (PLEG): container finished" podID="6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02" containerID="4b3ae247a77b9ca86a18b68baa10e2405c6aa764dbf08e992ffd9b53ab46525d" exitCode=0 Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.940153 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsxbw" event={"ID":"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02","Type":"ContainerDied","Data":"4b3ae247a77b9ca86a18b68baa10e2405c6aa764dbf08e992ffd9b53ab46525d"} Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.940188 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsxbw" event={"ID":"6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02","Type":"ContainerDied","Data":"5edb755d47ba8408734633a728a4b1083973c0db6644884852e0aea796a7530f"} Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.940209 4739 scope.go:117] "RemoveContainer" containerID="4b3ae247a77b9ca86a18b68baa10e2405c6aa764dbf08e992ffd9b53ab46525d" Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.940371 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vsxbw" Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.948783 4739 generic.go:334] "Generic (PLEG): container finished" podID="97c5fbe3-8871-4cbc-885a-d02783e3fc4c" containerID="911a305809350d8657e5588469d216b7928791dac03e33a4f8d08490c0136ae4" exitCode=0 Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.948810 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-48dnc" event={"ID":"97c5fbe3-8871-4cbc-885a-d02783e3fc4c","Type":"ContainerDied","Data":"911a305809350d8657e5588469d216b7928791dac03e33a4f8d08490c0136ae4"} Dec 01 16:12:02 crc kubenswrapper[4739]: I1201 16:12:02.979954 4739 scope.go:117] "RemoveContainer" containerID="42947a3ac64c5ad508fa2e43a0b6f7f2ba1d3d66dfdd477f2eb9a5cf432071f0" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.012026 4739 scope.go:117] "RemoveContainer" containerID="a3b9db03ccba1e5150a66bd57116db3b2554e5bab18d25d56486e216dd26f250" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.021355 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.021391 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md9j4\" (UniqueName: \"kubernetes.io/projected/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-kube-api-access-md9j4\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.043750 4739 scope.go:117] "RemoveContainer" containerID="4b3ae247a77b9ca86a18b68baa10e2405c6aa764dbf08e992ffd9b53ab46525d" Dec 01 16:12:03 crc kubenswrapper[4739]: E1201 16:12:03.044189 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b3ae247a77b9ca86a18b68baa10e2405c6aa764dbf08e992ffd9b53ab46525d\": container with ID starting with 4b3ae247a77b9ca86a18b68baa10e2405c6aa764dbf08e992ffd9b53ab46525d not found: ID does not exist" containerID="4b3ae247a77b9ca86a18b68baa10e2405c6aa764dbf08e992ffd9b53ab46525d" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.044229 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b3ae247a77b9ca86a18b68baa10e2405c6aa764dbf08e992ffd9b53ab46525d"} err="failed to get container status \"4b3ae247a77b9ca86a18b68baa10e2405c6aa764dbf08e992ffd9b53ab46525d\": rpc error: code = NotFound desc = could not find container \"4b3ae247a77b9ca86a18b68baa10e2405c6aa764dbf08e992ffd9b53ab46525d\": container with ID starting with 4b3ae247a77b9ca86a18b68baa10e2405c6aa764dbf08e992ffd9b53ab46525d not found: ID does not exist" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.044256 4739 scope.go:117] "RemoveContainer" containerID="42947a3ac64c5ad508fa2e43a0b6f7f2ba1d3d66dfdd477f2eb9a5cf432071f0" Dec 01 16:12:03 crc kubenswrapper[4739]: E1201 16:12:03.044700 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42947a3ac64c5ad508fa2e43a0b6f7f2ba1d3d66dfdd477f2eb9a5cf432071f0\": container with ID starting with 42947a3ac64c5ad508fa2e43a0b6f7f2ba1d3d66dfdd477f2eb9a5cf432071f0 not found: ID does not exist" containerID="42947a3ac64c5ad508fa2e43a0b6f7f2ba1d3d66dfdd477f2eb9a5cf432071f0" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.044748 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42947a3ac64c5ad508fa2e43a0b6f7f2ba1d3d66dfdd477f2eb9a5cf432071f0"} err="failed to get container status \"42947a3ac64c5ad508fa2e43a0b6f7f2ba1d3d66dfdd477f2eb9a5cf432071f0\": rpc error: code = NotFound desc = could not find container \"42947a3ac64c5ad508fa2e43a0b6f7f2ba1d3d66dfdd477f2eb9a5cf432071f0\": container with ID starting with 42947a3ac64c5ad508fa2e43a0b6f7f2ba1d3d66dfdd477f2eb9a5cf432071f0 not found: ID does not exist" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.044778 4739 scope.go:117] "RemoveContainer" containerID="a3b9db03ccba1e5150a66bd57116db3b2554e5bab18d25d56486e216dd26f250" Dec 01 16:12:03 crc kubenswrapper[4739]: E1201 16:12:03.045206 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3b9db03ccba1e5150a66bd57116db3b2554e5bab18d25d56486e216dd26f250\": container with ID starting with a3b9db03ccba1e5150a66bd57116db3b2554e5bab18d25d56486e216dd26f250 not found: ID does not exist" containerID="a3b9db03ccba1e5150a66bd57116db3b2554e5bab18d25d56486e216dd26f250" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.045238 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b9db03ccba1e5150a66bd57116db3b2554e5bab18d25d56486e216dd26f250"} err="failed to get container status \"a3b9db03ccba1e5150a66bd57116db3b2554e5bab18d25d56486e216dd26f250\": rpc error: code = NotFound desc = could not find container \"a3b9db03ccba1e5150a66bd57116db3b2554e5bab18d25d56486e216dd26f250\": container with ID starting with a3b9db03ccba1e5150a66bd57116db3b2554e5bab18d25d56486e216dd26f250 not found: ID does not exist" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.232910 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.321620 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02" (UID: "6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.325505 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-utilities\") pod \"97c5fbe3-8871-4cbc-885a-d02783e3fc4c\" (UID: \"97c5fbe3-8871-4cbc-885a-d02783e3fc4c\") " Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.325620 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-catalog-content\") pod \"97c5fbe3-8871-4cbc-885a-d02783e3fc4c\" (UID: \"97c5fbe3-8871-4cbc-885a-d02783e3fc4c\") " Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.325676 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhl9n\" (UniqueName: \"kubernetes.io/projected/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-kube-api-access-fhl9n\") pod \"97c5fbe3-8871-4cbc-885a-d02783e3fc4c\" (UID: \"97c5fbe3-8871-4cbc-885a-d02783e3fc4c\") " Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.326107 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.327356 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-utilities" (OuterVolumeSpecName: "utilities") pod "97c5fbe3-8871-4cbc-885a-d02783e3fc4c" (UID: "97c5fbe3-8871-4cbc-885a-d02783e3fc4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.333678 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-kube-api-access-fhl9n" (OuterVolumeSpecName: "kube-api-access-fhl9n") pod "97c5fbe3-8871-4cbc-885a-d02783e3fc4c" (UID: "97c5fbe3-8871-4cbc-885a-d02783e3fc4c"). InnerVolumeSpecName "kube-api-access-fhl9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.344527 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97c5fbe3-8871-4cbc-885a-d02783e3fc4c" (UID: "97c5fbe3-8871-4cbc-885a-d02783e3fc4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.428057 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.428122 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhl9n\" (UniqueName: \"kubernetes.io/projected/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-kube-api-access-fhl9n\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.428133 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c5fbe3-8871-4cbc-885a-d02783e3fc4c-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.583893 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vsxbw"] Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.592941 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vsxbw"] Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.959457 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-48dnc" event={"ID":"97c5fbe3-8871-4cbc-885a-d02783e3fc4c","Type":"ContainerDied","Data":"1ac069673522aefc254e07eea89578d34f214fe02f11438def83edc87af75de8"} Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.959499 4739 scope.go:117] "RemoveContainer" containerID="911a305809350d8657e5588469d216b7928791dac03e33a4f8d08490c0136ae4" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.959522 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-48dnc" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.981607 4739 scope.go:117] "RemoveContainer" containerID="4924ec711b09779cc99715dfef504021763fcec7440053380ca09676c71078ea" Dec 01 16:12:03 crc kubenswrapper[4739]: I1201 16:12:03.989926 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-48dnc"] Dec 01 16:12:04 crc kubenswrapper[4739]: I1201 16:12:04.002064 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-48dnc"] Dec 01 16:12:04 crc kubenswrapper[4739]: I1201 16:12:04.014071 4739 scope.go:117] "RemoveContainer" containerID="d75917330552ef8ac38f1bd2ebba3afbd7c0f8a8eb031459d4e841cf382c66ac" Dec 01 16:12:04 crc kubenswrapper[4739]: I1201 16:12:04.487300 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02" path="/var/lib/kubelet/pods/6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02/volumes" Dec 01 16:12:04 crc kubenswrapper[4739]: I1201 16:12:04.488322 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97c5fbe3-8871-4cbc-885a-d02783e3fc4c" path="/var/lib/kubelet/pods/97c5fbe3-8871-4cbc-885a-d02783e3fc4c/volumes" Dec 01 16:12:12 crc kubenswrapper[4739]: I1201 16:12:12.477389 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:12:12 crc kubenswrapper[4739]: E1201 16:12:12.478241 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:12:17 crc kubenswrapper[4739]: I1201 16:12:17.079989 4739 generic.go:334] "Generic (PLEG): container finished" podID="0c99055a-3919-465d-8584-a82a814a4be7" containerID="142a149e28f4080bf1114bdd56960de36aeb1a277453f032847583c620ed55c1" exitCode=0 Dec 01 16:12:17 crc kubenswrapper[4739]: I1201 16:12:17.080051 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" event={"ID":"0c99055a-3919-465d-8584-a82a814a4be7","Type":"ContainerDied","Data":"142a149e28f4080bf1114bdd56960de36aeb1a277453f032847583c620ed55c1"} Dec 01 16:12:18 crc kubenswrapper[4739]: I1201 16:12:18.552909 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:12:18 crc kubenswrapper[4739]: I1201 16:12:18.626120 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-ssh-key\") pod \"0c99055a-3919-465d-8584-a82a814a4be7\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " Dec 01 16:12:18 crc kubenswrapper[4739]: I1201 16:12:18.626221 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-ceph\") pod \"0c99055a-3919-465d-8584-a82a814a4be7\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " Dec 01 16:12:18 crc kubenswrapper[4739]: I1201 16:12:18.626246 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-inventory\") pod \"0c99055a-3919-465d-8584-a82a814a4be7\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " Dec 01 16:12:18 crc kubenswrapper[4739]: I1201 16:12:18.626347 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxs7b\" (UniqueName: \"kubernetes.io/projected/0c99055a-3919-465d-8584-a82a814a4be7-kube-api-access-jxs7b\") pod \"0c99055a-3919-465d-8584-a82a814a4be7\" (UID: \"0c99055a-3919-465d-8584-a82a814a4be7\") " Dec 01 16:12:18 crc kubenswrapper[4739]: I1201 16:12:18.631330 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-ceph" (OuterVolumeSpecName: "ceph") pod "0c99055a-3919-465d-8584-a82a814a4be7" (UID: "0c99055a-3919-465d-8584-a82a814a4be7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:18 crc kubenswrapper[4739]: I1201 16:12:18.632109 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c99055a-3919-465d-8584-a82a814a4be7-kube-api-access-jxs7b" (OuterVolumeSpecName: "kube-api-access-jxs7b") pod "0c99055a-3919-465d-8584-a82a814a4be7" (UID: "0c99055a-3919-465d-8584-a82a814a4be7"). InnerVolumeSpecName "kube-api-access-jxs7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:12:18 crc kubenswrapper[4739]: I1201 16:12:18.656044 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-inventory" (OuterVolumeSpecName: "inventory") pod "0c99055a-3919-465d-8584-a82a814a4be7" (UID: "0c99055a-3919-465d-8584-a82a814a4be7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:18 crc kubenswrapper[4739]: I1201 16:12:18.657797 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0c99055a-3919-465d-8584-a82a814a4be7" (UID: "0c99055a-3919-465d-8584-a82a814a4be7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:18 crc kubenswrapper[4739]: I1201 16:12:18.728076 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxs7b\" (UniqueName: \"kubernetes.io/projected/0c99055a-3919-465d-8584-a82a814a4be7-kube-api-access-jxs7b\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:18 crc kubenswrapper[4739]: I1201 16:12:18.728107 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:18 crc kubenswrapper[4739]: I1201 16:12:18.728119 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:18 crc kubenswrapper[4739]: I1201 16:12:18.728131 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c99055a-3919-465d-8584-a82a814a4be7-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.101269 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" event={"ID":"0c99055a-3919-465d-8584-a82a814a4be7","Type":"ContainerDied","Data":"d7c00ed0a39f8179b9b0cc533f093e3ae81060fe4c27f0dd785b3ff54a49bc1f"} Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.101326 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7c00ed0a39f8179b9b0cc533f093e3ae81060fe4c27f0dd785b3ff54a49bc1f" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.101356 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.196966 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk"] Dec 01 16:12:19 crc kubenswrapper[4739]: E1201 16:12:19.197516 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c99055a-3919-465d-8584-a82a814a4be7" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.197547 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c99055a-3919-465d-8584-a82a814a4be7" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 16:12:19 crc kubenswrapper[4739]: E1201 16:12:19.197564 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02" containerName="extract-utilities" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.197575 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02" containerName="extract-utilities" Dec 01 16:12:19 crc kubenswrapper[4739]: E1201 16:12:19.197591 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02" containerName="registry-server" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.197601 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02" containerName="registry-server" Dec 01 16:12:19 crc kubenswrapper[4739]: E1201 16:12:19.197623 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c5fbe3-8871-4cbc-885a-d02783e3fc4c" containerName="registry-server" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.197633 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c5fbe3-8871-4cbc-885a-d02783e3fc4c" containerName="registry-server" Dec 01 16:12:19 crc kubenswrapper[4739]: E1201 16:12:19.197657 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c5fbe3-8871-4cbc-885a-d02783e3fc4c" containerName="extract-utilities" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.197667 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c5fbe3-8871-4cbc-885a-d02783e3fc4c" containerName="extract-utilities" Dec 01 16:12:19 crc kubenswrapper[4739]: E1201 16:12:19.197684 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c5fbe3-8871-4cbc-885a-d02783e3fc4c" containerName="extract-content" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.197694 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c5fbe3-8871-4cbc-885a-d02783e3fc4c" containerName="extract-content" Dec 01 16:12:19 crc kubenswrapper[4739]: E1201 16:12:19.197712 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02" containerName="extract-content" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.197723 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02" containerName="extract-content" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.197972 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c99055a-3919-465d-8584-a82a814a4be7" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.197995 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d73bd1d-f9f6-4bef-8fe7-06f6a67b1a02" containerName="registry-server" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.198020 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="97c5fbe3-8871-4cbc-885a-d02783e3fc4c" containerName="registry-server" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.198928 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.201107 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.201849 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.203030 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.203301 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.203487 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.206538 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk"] Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.235103 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.235150 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.235204 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.235231 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4fbk\" (UniqueName: \"kubernetes.io/projected/ddf28305-5484-4d69-be7c-04fe79db3741-kube-api-access-p4fbk\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.337368 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.337595 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.337701 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.337789 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4fbk\" (UniqueName: \"kubernetes.io/projected/ddf28305-5484-4d69-be7c-04fe79db3741-kube-api-access-p4fbk\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.342796 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.344188 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.349862 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.358515 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4fbk\" (UniqueName: \"kubernetes.io/projected/ddf28305-5484-4d69-be7c-04fe79db3741-kube-api-access-p4fbk\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:19 crc kubenswrapper[4739]: I1201 16:12:19.535080 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:20 crc kubenswrapper[4739]: I1201 16:12:20.200024 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk"] Dec 01 16:12:21 crc kubenswrapper[4739]: I1201 16:12:21.131373 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" event={"ID":"ddf28305-5484-4d69-be7c-04fe79db3741","Type":"ContainerStarted","Data":"11d689155e0a197360cbc13bc0f0823015b8f235feb7e371300bad70c49cd1eb"} Dec 01 16:12:21 crc kubenswrapper[4739]: I1201 16:12:21.131676 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" event={"ID":"ddf28305-5484-4d69-be7c-04fe79db3741","Type":"ContainerStarted","Data":"c421e0932dd1b410b383346a1a6412ef7d339aa53b1387e9e8478ecc8938a0af"} Dec 01 16:12:21 crc kubenswrapper[4739]: I1201 16:12:21.158977 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" podStartSLOduration=1.630596342 podStartE2EDuration="2.158952477s" podCreationTimestamp="2025-12-01 16:12:19 +0000 UTC" firstStartedPulling="2025-12-01 16:12:20.206460878 +0000 UTC m=+2242.032206972" lastFinishedPulling="2025-12-01 16:12:20.734817013 +0000 UTC m=+2242.560563107" observedRunningTime="2025-12-01 16:12:21.155906342 +0000 UTC m=+2242.981652476" watchObservedRunningTime="2025-12-01 16:12:21.158952477 +0000 UTC m=+2242.984698571" Dec 01 16:12:23 crc kubenswrapper[4739]: I1201 16:12:23.477079 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:12:23 crc kubenswrapper[4739]: E1201 16:12:23.477724 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:12:26 crc kubenswrapper[4739]: I1201 16:12:26.181353 4739 generic.go:334] "Generic (PLEG): container finished" podID="ddf28305-5484-4d69-be7c-04fe79db3741" containerID="11d689155e0a197360cbc13bc0f0823015b8f235feb7e371300bad70c49cd1eb" exitCode=0 Dec 01 16:12:26 crc kubenswrapper[4739]: I1201 16:12:26.181470 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" event={"ID":"ddf28305-5484-4d69-be7c-04fe79db3741","Type":"ContainerDied","Data":"11d689155e0a197360cbc13bc0f0823015b8f235feb7e371300bad70c49cd1eb"} Dec 01 16:12:27 crc kubenswrapper[4739]: I1201 16:12:27.635478 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:27 crc kubenswrapper[4739]: I1201 16:12:27.822007 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4fbk\" (UniqueName: \"kubernetes.io/projected/ddf28305-5484-4d69-be7c-04fe79db3741-kube-api-access-p4fbk\") pod \"ddf28305-5484-4d69-be7c-04fe79db3741\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " Dec 01 16:12:27 crc kubenswrapper[4739]: I1201 16:12:27.822079 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-inventory\") pod \"ddf28305-5484-4d69-be7c-04fe79db3741\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " Dec 01 16:12:27 crc kubenswrapper[4739]: I1201 16:12:27.822144 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-ceph\") pod \"ddf28305-5484-4d69-be7c-04fe79db3741\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " Dec 01 16:12:27 crc kubenswrapper[4739]: I1201 16:12:27.822303 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-ssh-key\") pod \"ddf28305-5484-4d69-be7c-04fe79db3741\" (UID: \"ddf28305-5484-4d69-be7c-04fe79db3741\") " Dec 01 16:12:27 crc kubenswrapper[4739]: I1201 16:12:27.831181 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-ceph" (OuterVolumeSpecName: "ceph") pod "ddf28305-5484-4d69-be7c-04fe79db3741" (UID: "ddf28305-5484-4d69-be7c-04fe79db3741"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:27 crc kubenswrapper[4739]: I1201 16:12:27.840785 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddf28305-5484-4d69-be7c-04fe79db3741-kube-api-access-p4fbk" (OuterVolumeSpecName: "kube-api-access-p4fbk") pod "ddf28305-5484-4d69-be7c-04fe79db3741" (UID: "ddf28305-5484-4d69-be7c-04fe79db3741"). InnerVolumeSpecName "kube-api-access-p4fbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:12:27 crc kubenswrapper[4739]: I1201 16:12:27.859275 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ddf28305-5484-4d69-be7c-04fe79db3741" (UID: "ddf28305-5484-4d69-be7c-04fe79db3741"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:27 crc kubenswrapper[4739]: I1201 16:12:27.863837 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-inventory" (OuterVolumeSpecName: "inventory") pod "ddf28305-5484-4d69-be7c-04fe79db3741" (UID: "ddf28305-5484-4d69-be7c-04fe79db3741"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:27 crc kubenswrapper[4739]: I1201 16:12:27.924310 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:27 crc kubenswrapper[4739]: I1201 16:12:27.924344 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4fbk\" (UniqueName: \"kubernetes.io/projected/ddf28305-5484-4d69-be7c-04fe79db3741-kube-api-access-p4fbk\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:27 crc kubenswrapper[4739]: I1201 16:12:27.924356 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:27 crc kubenswrapper[4739]: I1201 16:12:27.924364 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ddf28305-5484-4d69-be7c-04fe79db3741-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.208751 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" event={"ID":"ddf28305-5484-4d69-be7c-04fe79db3741","Type":"ContainerDied","Data":"c421e0932dd1b410b383346a1a6412ef7d339aa53b1387e9e8478ecc8938a0af"} Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.208805 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c421e0932dd1b410b383346a1a6412ef7d339aa53b1387e9e8478ecc8938a0af" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.208894 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.315685 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm"] Dec 01 16:12:28 crc kubenswrapper[4739]: E1201 16:12:28.316100 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf28305-5484-4d69-be7c-04fe79db3741" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.316118 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf28305-5484-4d69-be7c-04fe79db3741" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.316339 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddf28305-5484-4d69-be7c-04fe79db3741" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.321138 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.324296 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.325100 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.325331 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.325512 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.326057 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.335925 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm"] Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.348324 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k9trm\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.348460 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k9trm\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.348505 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k9trm\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.348702 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwj84\" (UniqueName: \"kubernetes.io/projected/d28dec39-8ad1-4891-82bd-ce6080d608b8-kube-api-access-lwj84\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k9trm\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.450791 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k9trm\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.450838 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k9trm\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.450856 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k9trm\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.450910 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwj84\" (UniqueName: \"kubernetes.io/projected/d28dec39-8ad1-4891-82bd-ce6080d608b8-kube-api-access-lwj84\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k9trm\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.454728 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k9trm\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.454998 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k9trm\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.469569 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k9trm\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.474970 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwj84\" (UniqueName: \"kubernetes.io/projected/d28dec39-8ad1-4891-82bd-ce6080d608b8-kube-api-access-lwj84\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k9trm\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:12:28 crc kubenswrapper[4739]: I1201 16:12:28.649886 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:12:29 crc kubenswrapper[4739]: I1201 16:12:29.280394 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm"] Dec 01 16:12:30 crc kubenswrapper[4739]: I1201 16:12:30.232651 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" event={"ID":"d28dec39-8ad1-4891-82bd-ce6080d608b8","Type":"ContainerStarted","Data":"c3d82d873c0f802730d018e2c2c891ad41489c722c0e9a84940a74f973be8f8d"} Dec 01 16:12:30 crc kubenswrapper[4739]: I1201 16:12:30.232940 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" event={"ID":"d28dec39-8ad1-4891-82bd-ce6080d608b8","Type":"ContainerStarted","Data":"88093d765d8a997530d232e52e7c5e72e68b4c429d6bc4fbe802872e46892644"} Dec 01 16:12:30 crc kubenswrapper[4739]: I1201 16:12:30.252164 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" podStartSLOduration=1.598541005 podStartE2EDuration="2.252141519s" podCreationTimestamp="2025-12-01 16:12:28 +0000 UTC" firstStartedPulling="2025-12-01 16:12:29.280620253 +0000 UTC m=+2251.106366387" lastFinishedPulling="2025-12-01 16:12:29.934220787 +0000 UTC m=+2251.759966901" observedRunningTime="2025-12-01 16:12:30.249641512 +0000 UTC m=+2252.075387606" watchObservedRunningTime="2025-12-01 16:12:30.252141519 +0000 UTC m=+2252.077887623" Dec 01 16:12:38 crc kubenswrapper[4739]: I1201 16:12:38.492694 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:12:38 crc kubenswrapper[4739]: E1201 16:12:38.495221 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:12:51 crc kubenswrapper[4739]: I1201 16:12:51.477807 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:12:51 crc kubenswrapper[4739]: E1201 16:12:51.478672 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:13:05 crc kubenswrapper[4739]: I1201 16:13:05.477465 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:13:05 crc kubenswrapper[4739]: E1201 16:13:05.478555 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:13:10 crc kubenswrapper[4739]: I1201 16:13:10.604334 4739 generic.go:334] "Generic (PLEG): container finished" podID="d28dec39-8ad1-4891-82bd-ce6080d608b8" containerID="c3d82d873c0f802730d018e2c2c891ad41489c722c0e9a84940a74f973be8f8d" exitCode=0 Dec 01 16:13:10 crc kubenswrapper[4739]: I1201 16:13:10.604454 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" event={"ID":"d28dec39-8ad1-4891-82bd-ce6080d608b8","Type":"ContainerDied","Data":"c3d82d873c0f802730d018e2c2c891ad41489c722c0e9a84940a74f973be8f8d"} Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.093612 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.189637 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-ceph\") pod \"d28dec39-8ad1-4891-82bd-ce6080d608b8\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.189832 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwj84\" (UniqueName: \"kubernetes.io/projected/d28dec39-8ad1-4891-82bd-ce6080d608b8-kube-api-access-lwj84\") pod \"d28dec39-8ad1-4891-82bd-ce6080d608b8\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.189881 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-ssh-key\") pod \"d28dec39-8ad1-4891-82bd-ce6080d608b8\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.189984 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-inventory\") pod \"d28dec39-8ad1-4891-82bd-ce6080d608b8\" (UID: \"d28dec39-8ad1-4891-82bd-ce6080d608b8\") " Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.200542 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-ceph" (OuterVolumeSpecName: "ceph") pod "d28dec39-8ad1-4891-82bd-ce6080d608b8" (UID: "d28dec39-8ad1-4891-82bd-ce6080d608b8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.200968 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d28dec39-8ad1-4891-82bd-ce6080d608b8-kube-api-access-lwj84" (OuterVolumeSpecName: "kube-api-access-lwj84") pod "d28dec39-8ad1-4891-82bd-ce6080d608b8" (UID: "d28dec39-8ad1-4891-82bd-ce6080d608b8"). InnerVolumeSpecName "kube-api-access-lwj84". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.232596 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-inventory" (OuterVolumeSpecName: "inventory") pod "d28dec39-8ad1-4891-82bd-ce6080d608b8" (UID: "d28dec39-8ad1-4891-82bd-ce6080d608b8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.270106 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d28dec39-8ad1-4891-82bd-ce6080d608b8" (UID: "d28dec39-8ad1-4891-82bd-ce6080d608b8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.297997 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwj84\" (UniqueName: \"kubernetes.io/projected/d28dec39-8ad1-4891-82bd-ce6080d608b8-kube-api-access-lwj84\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.298039 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.298052 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.298063 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d28dec39-8ad1-4891-82bd-ce6080d608b8-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.625104 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" event={"ID":"d28dec39-8ad1-4891-82bd-ce6080d608b8","Type":"ContainerDied","Data":"88093d765d8a997530d232e52e7c5e72e68b4c429d6bc4fbe802872e46892644"} Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.625447 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88093d765d8a997530d232e52e7c5e72e68b4c429d6bc4fbe802872e46892644" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.625191 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k9trm" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.716122 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7"] Dec 01 16:13:12 crc kubenswrapper[4739]: E1201 16:13:12.717229 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d28dec39-8ad1-4891-82bd-ce6080d608b8" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.717479 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="d28dec39-8ad1-4891-82bd-ce6080d608b8" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.718070 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="d28dec39-8ad1-4891-82bd-ce6080d608b8" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.719237 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.722389 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.723093 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.724006 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.724207 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.724522 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7"] Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.725736 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.807401 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.807553 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkpzs\" (UniqueName: \"kubernetes.io/projected/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-kube-api-access-kkpzs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.807687 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.807809 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.909145 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkpzs\" (UniqueName: \"kubernetes.io/projected/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-kube-api-access-kkpzs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.909244 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.909340 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.909401 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.913981 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.915570 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.916562 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:12 crc kubenswrapper[4739]: I1201 16:13:12.939060 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkpzs\" (UniqueName: \"kubernetes.io/projected/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-kube-api-access-kkpzs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:13 crc kubenswrapper[4739]: I1201 16:13:13.043684 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:13 crc kubenswrapper[4739]: I1201 16:13:13.653798 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7"] Dec 01 16:13:14 crc kubenswrapper[4739]: I1201 16:13:14.647242 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" event={"ID":"af3127e6-d90c-4ac8-9e40-d8710d4f4bca","Type":"ContainerStarted","Data":"f3115e6625f34f689d5067f5d507388bf8da20e1fe17d6b147d02079ffae8a0f"} Dec 01 16:13:14 crc kubenswrapper[4739]: I1201 16:13:14.647665 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" event={"ID":"af3127e6-d90c-4ac8-9e40-d8710d4f4bca","Type":"ContainerStarted","Data":"b62aa3927a28ac2d1859d061ce7d3de2f668c45f1a75a9cd2d254bed3c1d472c"} Dec 01 16:13:14 crc kubenswrapper[4739]: I1201 16:13:14.676496 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" podStartSLOduration=2.120329189 podStartE2EDuration="2.676470571s" podCreationTimestamp="2025-12-01 16:13:12 +0000 UTC" firstStartedPulling="2025-12-01 16:13:13.655232959 +0000 UTC m=+2295.480979053" lastFinishedPulling="2025-12-01 16:13:14.211374301 +0000 UTC m=+2296.037120435" observedRunningTime="2025-12-01 16:13:14.66865741 +0000 UTC m=+2296.494403514" watchObservedRunningTime="2025-12-01 16:13:14.676470571 +0000 UTC m=+2296.502216705" Dec 01 16:13:16 crc kubenswrapper[4739]: I1201 16:13:16.477231 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:13:16 crc kubenswrapper[4739]: E1201 16:13:16.477638 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:13:18 crc kubenswrapper[4739]: I1201 16:13:18.721642 4739 generic.go:334] "Generic (PLEG): container finished" podID="af3127e6-d90c-4ac8-9e40-d8710d4f4bca" containerID="f3115e6625f34f689d5067f5d507388bf8da20e1fe17d6b147d02079ffae8a0f" exitCode=0 Dec 01 16:13:18 crc kubenswrapper[4739]: I1201 16:13:18.721749 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" event={"ID":"af3127e6-d90c-4ac8-9e40-d8710d4f4bca","Type":"ContainerDied","Data":"f3115e6625f34f689d5067f5d507388bf8da20e1fe17d6b147d02079ffae8a0f"} Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.214934 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.359845 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkpzs\" (UniqueName: \"kubernetes.io/projected/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-kube-api-access-kkpzs\") pod \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.359910 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-ceph\") pod \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.360034 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-inventory\") pod \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.360203 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-ssh-key\") pod \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\" (UID: \"af3127e6-d90c-4ac8-9e40-d8710d4f4bca\") " Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.365680 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-ceph" (OuterVolumeSpecName: "ceph") pod "af3127e6-d90c-4ac8-9e40-d8710d4f4bca" (UID: "af3127e6-d90c-4ac8-9e40-d8710d4f4bca"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.373648 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-kube-api-access-kkpzs" (OuterVolumeSpecName: "kube-api-access-kkpzs") pod "af3127e6-d90c-4ac8-9e40-d8710d4f4bca" (UID: "af3127e6-d90c-4ac8-9e40-d8710d4f4bca"). InnerVolumeSpecName "kube-api-access-kkpzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.387607 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-inventory" (OuterVolumeSpecName: "inventory") pod "af3127e6-d90c-4ac8-9e40-d8710d4f4bca" (UID: "af3127e6-d90c-4ac8-9e40-d8710d4f4bca"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.404109 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "af3127e6-d90c-4ac8-9e40-d8710d4f4bca" (UID: "af3127e6-d90c-4ac8-9e40-d8710d4f4bca"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.463916 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.463967 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.463991 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkpzs\" (UniqueName: \"kubernetes.io/projected/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-kube-api-access-kkpzs\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.464013 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/af3127e6-d90c-4ac8-9e40-d8710d4f4bca-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.742215 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" event={"ID":"af3127e6-d90c-4ac8-9e40-d8710d4f4bca","Type":"ContainerDied","Data":"b62aa3927a28ac2d1859d061ce7d3de2f668c45f1a75a9cd2d254bed3c1d472c"} Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.742286 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b62aa3927a28ac2d1859d061ce7d3de2f668c45f1a75a9cd2d254bed3c1d472c" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.742280 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.898726 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs"] Dec 01 16:13:20 crc kubenswrapper[4739]: E1201 16:13:20.899928 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3127e6-d90c-4ac8-9e40-d8710d4f4bca" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.899975 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3127e6-d90c-4ac8-9e40-d8710d4f4bca" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.900387 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3127e6-d90c-4ac8-9e40-d8710d4f4bca" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.901847 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.905126 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.905469 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.906361 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.906435 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.906594 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.913216 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs"] Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.975670 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfsv5\" (UniqueName: \"kubernetes.io/projected/7881548e-3e40-4ffc-85b5-d64050ea26c6-kube-api-access-hfsv5\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.975807 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.975845 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:13:20 crc kubenswrapper[4739]: I1201 16:13:20.976000 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:13:21 crc kubenswrapper[4739]: I1201 16:13:21.078183 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:13:21 crc kubenswrapper[4739]: I1201 16:13:21.078283 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfsv5\" (UniqueName: \"kubernetes.io/projected/7881548e-3e40-4ffc-85b5-d64050ea26c6-kube-api-access-hfsv5\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:13:21 crc kubenswrapper[4739]: I1201 16:13:21.078373 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:13:21 crc kubenswrapper[4739]: I1201 16:13:21.078403 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:13:21 crc kubenswrapper[4739]: I1201 16:13:21.089187 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:13:21 crc kubenswrapper[4739]: I1201 16:13:21.089250 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:13:21 crc kubenswrapper[4739]: I1201 16:13:21.092313 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:13:21 crc kubenswrapper[4739]: I1201 16:13:21.165015 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfsv5\" (UniqueName: \"kubernetes.io/projected/7881548e-3e40-4ffc-85b5-d64050ea26c6-kube-api-access-hfsv5\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:13:21 crc kubenswrapper[4739]: I1201 16:13:21.226730 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:13:21 crc kubenswrapper[4739]: I1201 16:13:21.772516 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs"] Dec 01 16:13:22 crc kubenswrapper[4739]: I1201 16:13:22.774373 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" event={"ID":"7881548e-3e40-4ffc-85b5-d64050ea26c6","Type":"ContainerStarted","Data":"c66c9f0ddefb117bf4c1f4901c1818bb978387a568cd918057179769ec48a97b"} Dec 01 16:13:22 crc kubenswrapper[4739]: I1201 16:13:22.777110 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" event={"ID":"7881548e-3e40-4ffc-85b5-d64050ea26c6","Type":"ContainerStarted","Data":"487bfe40706e80c12ea311baabcc96e96aaeaf103bada587b49f4e2ab9c9b742"} Dec 01 16:13:22 crc kubenswrapper[4739]: I1201 16:13:22.807909 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" podStartSLOduration=2.234813061 podStartE2EDuration="2.807877297s" podCreationTimestamp="2025-12-01 16:13:20 +0000 UTC" firstStartedPulling="2025-12-01 16:13:21.78290603 +0000 UTC m=+2303.608652134" lastFinishedPulling="2025-12-01 16:13:22.355970236 +0000 UTC m=+2304.181716370" observedRunningTime="2025-12-01 16:13:22.80114364 +0000 UTC m=+2304.626889764" watchObservedRunningTime="2025-12-01 16:13:22.807877297 +0000 UTC m=+2304.633623431" Dec 01 16:13:30 crc kubenswrapper[4739]: I1201 16:13:30.477918 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:13:30 crc kubenswrapper[4739]: E1201 16:13:30.478739 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:13:44 crc kubenswrapper[4739]: I1201 16:13:44.478335 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:13:44 crc kubenswrapper[4739]: E1201 16:13:44.479901 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:13:58 crc kubenswrapper[4739]: I1201 16:13:58.487739 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:13:58 crc kubenswrapper[4739]: E1201 16:13:58.489023 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:14:11 crc kubenswrapper[4739]: I1201 16:14:11.657547 4739 generic.go:334] "Generic (PLEG): container finished" podID="7881548e-3e40-4ffc-85b5-d64050ea26c6" containerID="c66c9f0ddefb117bf4c1f4901c1818bb978387a568cd918057179769ec48a97b" exitCode=0 Dec 01 16:14:11 crc kubenswrapper[4739]: I1201 16:14:11.657757 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" event={"ID":"7881548e-3e40-4ffc-85b5-d64050ea26c6","Type":"ContainerDied","Data":"c66c9f0ddefb117bf4c1f4901c1818bb978387a568cd918057179769ec48a97b"} Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.197929 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.397957 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfsv5\" (UniqueName: \"kubernetes.io/projected/7881548e-3e40-4ffc-85b5-d64050ea26c6-kube-api-access-hfsv5\") pod \"7881548e-3e40-4ffc-85b5-d64050ea26c6\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.398087 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-inventory\") pod \"7881548e-3e40-4ffc-85b5-d64050ea26c6\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.399574 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-ceph\") pod \"7881548e-3e40-4ffc-85b5-d64050ea26c6\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.399723 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-ssh-key\") pod \"7881548e-3e40-4ffc-85b5-d64050ea26c6\" (UID: \"7881548e-3e40-4ffc-85b5-d64050ea26c6\") " Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.405706 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-ceph" (OuterVolumeSpecName: "ceph") pod "7881548e-3e40-4ffc-85b5-d64050ea26c6" (UID: "7881548e-3e40-4ffc-85b5-d64050ea26c6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.408788 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7881548e-3e40-4ffc-85b5-d64050ea26c6-kube-api-access-hfsv5" (OuterVolumeSpecName: "kube-api-access-hfsv5") pod "7881548e-3e40-4ffc-85b5-d64050ea26c6" (UID: "7881548e-3e40-4ffc-85b5-d64050ea26c6"). InnerVolumeSpecName "kube-api-access-hfsv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.431940 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-inventory" (OuterVolumeSpecName: "inventory") pod "7881548e-3e40-4ffc-85b5-d64050ea26c6" (UID: "7881548e-3e40-4ffc-85b5-d64050ea26c6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.445697 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7881548e-3e40-4ffc-85b5-d64050ea26c6" (UID: "7881548e-3e40-4ffc-85b5-d64050ea26c6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.478349 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:14:13 crc kubenswrapper[4739]: E1201 16:14:13.479010 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.505033 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfsv5\" (UniqueName: \"kubernetes.io/projected/7881548e-3e40-4ffc-85b5-d64050ea26c6-kube-api-access-hfsv5\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.505332 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.505612 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.505861 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7881548e-3e40-4ffc-85b5-d64050ea26c6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.681529 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" event={"ID":"7881548e-3e40-4ffc-85b5-d64050ea26c6","Type":"ContainerDied","Data":"487bfe40706e80c12ea311baabcc96e96aaeaf103bada587b49f4e2ab9c9b742"} Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.681600 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="487bfe40706e80c12ea311baabcc96e96aaeaf103bada587b49f4e2ab9c9b742" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.681624 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.774651 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-vq2hk"] Dec 01 16:14:13 crc kubenswrapper[4739]: E1201 16:14:13.775195 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7881548e-3e40-4ffc-85b5-d64050ea26c6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.775217 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7881548e-3e40-4ffc-85b5-d64050ea26c6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.775473 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="7881548e-3e40-4ffc-85b5-d64050ea26c6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.776297 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.778466 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.779151 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.779305 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.779352 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.784119 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.785830 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-vq2hk"] Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.819969 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-ceph\") pod \"ssh-known-hosts-edpm-deployment-vq2hk\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.820051 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l528\" (UniqueName: \"kubernetes.io/projected/b45dd823-9e52-42f9-8bcf-5b27757f46e2-kube-api-access-2l528\") pod \"ssh-known-hosts-edpm-deployment-vq2hk\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.820104 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-vq2hk\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.820532 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-vq2hk\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.921960 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-vq2hk\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.922036 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-ceph\") pod \"ssh-known-hosts-edpm-deployment-vq2hk\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.922070 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l528\" (UniqueName: \"kubernetes.io/projected/b45dd823-9e52-42f9-8bcf-5b27757f46e2-kube-api-access-2l528\") pod \"ssh-known-hosts-edpm-deployment-vq2hk\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.922112 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-vq2hk\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.927124 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-ceph\") pod \"ssh-known-hosts-edpm-deployment-vq2hk\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.927884 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-vq2hk\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.931744 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-vq2hk\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:13 crc kubenswrapper[4739]: I1201 16:14:13.965222 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l528\" (UniqueName: \"kubernetes.io/projected/b45dd823-9e52-42f9-8bcf-5b27757f46e2-kube-api-access-2l528\") pod \"ssh-known-hosts-edpm-deployment-vq2hk\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:14 crc kubenswrapper[4739]: I1201 16:14:14.093933 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:14 crc kubenswrapper[4739]: I1201 16:14:14.677644 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-vq2hk"] Dec 01 16:14:14 crc kubenswrapper[4739]: W1201 16:14:14.691683 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb45dd823_9e52_42f9_8bcf_5b27757f46e2.slice/crio-1aa06f998a388981665c2ddae7dfc13e91c6d1e9062da9759dfaa88b30e21414 WatchSource:0}: Error finding container 1aa06f998a388981665c2ddae7dfc13e91c6d1e9062da9759dfaa88b30e21414: Status 404 returned error can't find the container with id 1aa06f998a388981665c2ddae7dfc13e91c6d1e9062da9759dfaa88b30e21414 Dec 01 16:14:14 crc kubenswrapper[4739]: I1201 16:14:14.694393 4739 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:14:15 crc kubenswrapper[4739]: I1201 16:14:15.697009 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" event={"ID":"b45dd823-9e52-42f9-8bcf-5b27757f46e2","Type":"ContainerStarted","Data":"1aa06f998a388981665c2ddae7dfc13e91c6d1e9062da9759dfaa88b30e21414"} Dec 01 16:14:16 crc kubenswrapper[4739]: I1201 16:14:16.707170 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" event={"ID":"b45dd823-9e52-42f9-8bcf-5b27757f46e2","Type":"ContainerStarted","Data":"6b094ff6cd239d96c42c3f3cec072e154eea9b2ae6497a9682a97b3e812e7d1b"} Dec 01 16:14:16 crc kubenswrapper[4739]: I1201 16:14:16.726335 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" podStartSLOduration=2.855459691 podStartE2EDuration="3.726315521s" podCreationTimestamp="2025-12-01 16:14:13 +0000 UTC" firstStartedPulling="2025-12-01 16:14:14.69422348 +0000 UTC m=+2356.519969574" lastFinishedPulling="2025-12-01 16:14:15.56507931 +0000 UTC m=+2357.390825404" observedRunningTime="2025-12-01 16:14:16.720729008 +0000 UTC m=+2358.546475102" watchObservedRunningTime="2025-12-01 16:14:16.726315521 +0000 UTC m=+2358.552061605" Dec 01 16:14:25 crc kubenswrapper[4739]: I1201 16:14:25.478211 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:14:25 crc kubenswrapper[4739]: E1201 16:14:25.479199 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:14:25 crc kubenswrapper[4739]: I1201 16:14:25.805962 4739 generic.go:334] "Generic (PLEG): container finished" podID="b45dd823-9e52-42f9-8bcf-5b27757f46e2" containerID="6b094ff6cd239d96c42c3f3cec072e154eea9b2ae6497a9682a97b3e812e7d1b" exitCode=0 Dec 01 16:14:25 crc kubenswrapper[4739]: I1201 16:14:25.806208 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" event={"ID":"b45dd823-9e52-42f9-8bcf-5b27757f46e2","Type":"ContainerDied","Data":"6b094ff6cd239d96c42c3f3cec072e154eea9b2ae6497a9682a97b3e812e7d1b"} Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.229228 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.380913 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-ceph\") pod \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.381169 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2l528\" (UniqueName: \"kubernetes.io/projected/b45dd823-9e52-42f9-8bcf-5b27757f46e2-kube-api-access-2l528\") pod \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.381300 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-inventory-0\") pod \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.381387 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-ssh-key-openstack-edpm-ipam\") pod \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\" (UID: \"b45dd823-9e52-42f9-8bcf-5b27757f46e2\") " Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.387769 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b45dd823-9e52-42f9-8bcf-5b27757f46e2-kube-api-access-2l528" (OuterVolumeSpecName: "kube-api-access-2l528") pod "b45dd823-9e52-42f9-8bcf-5b27757f46e2" (UID: "b45dd823-9e52-42f9-8bcf-5b27757f46e2"). InnerVolumeSpecName "kube-api-access-2l528". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.389574 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-ceph" (OuterVolumeSpecName: "ceph") pod "b45dd823-9e52-42f9-8bcf-5b27757f46e2" (UID: "b45dd823-9e52-42f9-8bcf-5b27757f46e2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.419032 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "b45dd823-9e52-42f9-8bcf-5b27757f46e2" (UID: "b45dd823-9e52-42f9-8bcf-5b27757f46e2"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.435615 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b45dd823-9e52-42f9-8bcf-5b27757f46e2" (UID: "b45dd823-9e52-42f9-8bcf-5b27757f46e2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.484641 4739 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.484774 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.484817 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b45dd823-9e52-42f9-8bcf-5b27757f46e2-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.484836 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2l528\" (UniqueName: \"kubernetes.io/projected/b45dd823-9e52-42f9-8bcf-5b27757f46e2-kube-api-access-2l528\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.834111 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" event={"ID":"b45dd823-9e52-42f9-8bcf-5b27757f46e2","Type":"ContainerDied","Data":"1aa06f998a388981665c2ddae7dfc13e91c6d1e9062da9759dfaa88b30e21414"} Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.834165 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1aa06f998a388981665c2ddae7dfc13e91c6d1e9062da9759dfaa88b30e21414" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.834192 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vq2hk" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.926940 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs"] Dec 01 16:14:27 crc kubenswrapper[4739]: E1201 16:14:27.927349 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b45dd823-9e52-42f9-8bcf-5b27757f46e2" containerName="ssh-known-hosts-edpm-deployment" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.927407 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="b45dd823-9e52-42f9-8bcf-5b27757f46e2" containerName="ssh-known-hosts-edpm-deployment" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.927655 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="b45dd823-9e52-42f9-8bcf-5b27757f46e2" containerName="ssh-known-hosts-edpm-deployment" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.928354 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.931222 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.931635 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.931966 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.932236 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.932471 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.949027 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs"] Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.998940 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wstvs\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.999347 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrfp6\" (UniqueName: \"kubernetes.io/projected/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-kube-api-access-hrfp6\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wstvs\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.999513 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wstvs\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:27 crc kubenswrapper[4739]: I1201 16:14:27.999979 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wstvs\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:28 crc kubenswrapper[4739]: I1201 16:14:28.102120 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrfp6\" (UniqueName: \"kubernetes.io/projected/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-kube-api-access-hrfp6\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wstvs\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:28 crc kubenswrapper[4739]: I1201 16:14:28.102185 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wstvs\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:28 crc kubenswrapper[4739]: I1201 16:14:28.102285 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wstvs\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:28 crc kubenswrapper[4739]: I1201 16:14:28.102355 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wstvs\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:28 crc kubenswrapper[4739]: I1201 16:14:28.110582 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wstvs\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:28 crc kubenswrapper[4739]: I1201 16:14:28.112214 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wstvs\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:28 crc kubenswrapper[4739]: I1201 16:14:28.122048 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wstvs\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:28 crc kubenswrapper[4739]: I1201 16:14:28.137548 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrfp6\" (UniqueName: \"kubernetes.io/projected/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-kube-api-access-hrfp6\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wstvs\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:28 crc kubenswrapper[4739]: I1201 16:14:28.252831 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:28 crc kubenswrapper[4739]: I1201 16:14:28.688432 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs"] Dec 01 16:14:28 crc kubenswrapper[4739]: I1201 16:14:28.845701 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" event={"ID":"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a","Type":"ContainerStarted","Data":"22cccfda0fa1ca39239240d055c714c240c56e20fe10453e00f6fc2723b134d8"} Dec 01 16:14:29 crc kubenswrapper[4739]: I1201 16:14:29.855330 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" event={"ID":"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a","Type":"ContainerStarted","Data":"dd19076df270fd8e8c45cfc351a188aefca6ba6fc3ffeb4ba9c8f5252d6d5aae"} Dec 01 16:14:29 crc kubenswrapper[4739]: I1201 16:14:29.881893 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" podStartSLOduration=2.197300953 podStartE2EDuration="2.88187375s" podCreationTimestamp="2025-12-01 16:14:27 +0000 UTC" firstStartedPulling="2025-12-01 16:14:28.700667632 +0000 UTC m=+2370.526413726" lastFinishedPulling="2025-12-01 16:14:29.385240389 +0000 UTC m=+2371.210986523" observedRunningTime="2025-12-01 16:14:29.877113352 +0000 UTC m=+2371.702859496" watchObservedRunningTime="2025-12-01 16:14:29.88187375 +0000 UTC m=+2371.707619864" Dec 01 16:14:37 crc kubenswrapper[4739]: I1201 16:14:37.932360 4739 generic.go:334] "Generic (PLEG): container finished" podID="d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a" containerID="dd19076df270fd8e8c45cfc351a188aefca6ba6fc3ffeb4ba9c8f5252d6d5aae" exitCode=0 Dec 01 16:14:37 crc kubenswrapper[4739]: I1201 16:14:37.932477 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" event={"ID":"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a","Type":"ContainerDied","Data":"dd19076df270fd8e8c45cfc351a188aefca6ba6fc3ffeb4ba9c8f5252d6d5aae"} Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.390531 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.529192 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-ssh-key\") pod \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.529322 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-inventory\") pod \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.529391 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-ceph\") pod \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.529469 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrfp6\" (UniqueName: \"kubernetes.io/projected/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-kube-api-access-hrfp6\") pod \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\" (UID: \"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a\") " Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.535175 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-ceph" (OuterVolumeSpecName: "ceph") pod "d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a" (UID: "d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.536196 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-kube-api-access-hrfp6" (OuterVolumeSpecName: "kube-api-access-hrfp6") pod "d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a" (UID: "d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a"). InnerVolumeSpecName "kube-api-access-hrfp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.553783 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-inventory" (OuterVolumeSpecName: "inventory") pod "d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a" (UID: "d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.574736 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a" (UID: "d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.631980 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.632036 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.632057 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.632076 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrfp6\" (UniqueName: \"kubernetes.io/projected/d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a-kube-api-access-hrfp6\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.952625 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" event={"ID":"d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a","Type":"ContainerDied","Data":"22cccfda0fa1ca39239240d055c714c240c56e20fe10453e00f6fc2723b134d8"} Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.952989 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22cccfda0fa1ca39239240d055c714c240c56e20fe10453e00f6fc2723b134d8" Dec 01 16:14:39 crc kubenswrapper[4739]: I1201 16:14:39.952764 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wstvs" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.047825 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8"] Dec 01 16:14:40 crc kubenswrapper[4739]: E1201 16:14:40.048155 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.048172 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.048366 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.048899 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.050820 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.051615 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.054269 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.054270 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.056575 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.068719 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8"] Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.143905 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.143973 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw7bk\" (UniqueName: \"kubernetes.io/projected/f111abd9-3325-4578-8429-f107f9ba3185-kube-api-access-cw7bk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.144059 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.144114 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.245265 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.245846 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.246337 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw7bk\" (UniqueName: \"kubernetes.io/projected/f111abd9-3325-4578-8429-f107f9ba3185-kube-api-access-cw7bk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.247036 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.249666 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.250124 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.250722 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.261121 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw7bk\" (UniqueName: \"kubernetes.io/projected/f111abd9-3325-4578-8429-f107f9ba3185-kube-api-access-cw7bk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.364958 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.491158 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:14:40 crc kubenswrapper[4739]: E1201 16:14:40.491651 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.903511 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8"] Dec 01 16:14:40 crc kubenswrapper[4739]: I1201 16:14:40.963570 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" event={"ID":"f111abd9-3325-4578-8429-f107f9ba3185","Type":"ContainerStarted","Data":"b3c800a7446550ee9dcfb30c9197faf9039cd4bc965412b8ee5e14fb220a96b4"} Dec 01 16:14:41 crc kubenswrapper[4739]: I1201 16:14:41.975256 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" event={"ID":"f111abd9-3325-4578-8429-f107f9ba3185","Type":"ContainerStarted","Data":"8f35afbb03cac4fee70ddd9fb902ed5a8401a824a570c5f4059e215e138f3edf"} Dec 01 16:14:42 crc kubenswrapper[4739]: I1201 16:14:42.015241 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" podStartSLOduration=1.404282472 podStartE2EDuration="2.015211503s" podCreationTimestamp="2025-12-01 16:14:40 +0000 UTC" firstStartedPulling="2025-12-01 16:14:40.912027186 +0000 UTC m=+2382.737773290" lastFinishedPulling="2025-12-01 16:14:41.522956207 +0000 UTC m=+2383.348702321" observedRunningTime="2025-12-01 16:14:42.001372586 +0000 UTC m=+2383.827118690" watchObservedRunningTime="2025-12-01 16:14:42.015211503 +0000 UTC m=+2383.840957617" Dec 01 16:14:52 crc kubenswrapper[4739]: I1201 16:14:52.098212 4739 generic.go:334] "Generic (PLEG): container finished" podID="f111abd9-3325-4578-8429-f107f9ba3185" containerID="8f35afbb03cac4fee70ddd9fb902ed5a8401a824a570c5f4059e215e138f3edf" exitCode=0 Dec 01 16:14:52 crc kubenswrapper[4739]: I1201 16:14:52.098849 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" event={"ID":"f111abd9-3325-4578-8429-f107f9ba3185","Type":"ContainerDied","Data":"8f35afbb03cac4fee70ddd9fb902ed5a8401a824a570c5f4059e215e138f3edf"} Dec 01 16:14:53 crc kubenswrapper[4739]: I1201 16:14:53.585075 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:53 crc kubenswrapper[4739]: I1201 16:14:53.741845 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-ceph\") pod \"f111abd9-3325-4578-8429-f107f9ba3185\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " Dec 01 16:14:53 crc kubenswrapper[4739]: I1201 16:14:53.742301 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw7bk\" (UniqueName: \"kubernetes.io/projected/f111abd9-3325-4578-8429-f107f9ba3185-kube-api-access-cw7bk\") pod \"f111abd9-3325-4578-8429-f107f9ba3185\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " Dec 01 16:14:53 crc kubenswrapper[4739]: I1201 16:14:53.742343 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-inventory\") pod \"f111abd9-3325-4578-8429-f107f9ba3185\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " Dec 01 16:14:53 crc kubenswrapper[4739]: I1201 16:14:53.742634 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-ssh-key\") pod \"f111abd9-3325-4578-8429-f107f9ba3185\" (UID: \"f111abd9-3325-4578-8429-f107f9ba3185\") " Dec 01 16:14:53 crc kubenswrapper[4739]: I1201 16:14:53.747673 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f111abd9-3325-4578-8429-f107f9ba3185-kube-api-access-cw7bk" (OuterVolumeSpecName: "kube-api-access-cw7bk") pod "f111abd9-3325-4578-8429-f107f9ba3185" (UID: "f111abd9-3325-4578-8429-f107f9ba3185"). InnerVolumeSpecName "kube-api-access-cw7bk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:14:53 crc kubenswrapper[4739]: I1201 16:14:53.754722 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-ceph" (OuterVolumeSpecName: "ceph") pod "f111abd9-3325-4578-8429-f107f9ba3185" (UID: "f111abd9-3325-4578-8429-f107f9ba3185"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:14:53 crc kubenswrapper[4739]: I1201 16:14:53.780685 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f111abd9-3325-4578-8429-f107f9ba3185" (UID: "f111abd9-3325-4578-8429-f107f9ba3185"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:14:53 crc kubenswrapper[4739]: I1201 16:14:53.783893 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-inventory" (OuterVolumeSpecName: "inventory") pod "f111abd9-3325-4578-8429-f107f9ba3185" (UID: "f111abd9-3325-4578-8429-f107f9ba3185"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:14:53 crc kubenswrapper[4739]: I1201 16:14:53.844541 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw7bk\" (UniqueName: \"kubernetes.io/projected/f111abd9-3325-4578-8429-f107f9ba3185-kube-api-access-cw7bk\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:53 crc kubenswrapper[4739]: I1201 16:14:53.844586 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:53 crc kubenswrapper[4739]: I1201 16:14:53.844600 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:53 crc kubenswrapper[4739]: I1201 16:14:53.844611 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f111abd9-3325-4578-8429-f107f9ba3185-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.121732 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" event={"ID":"f111abd9-3325-4578-8429-f107f9ba3185","Type":"ContainerDied","Data":"b3c800a7446550ee9dcfb30c9197faf9039cd4bc965412b8ee5e14fb220a96b4"} Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.121775 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3c800a7446550ee9dcfb30c9197faf9039cd4bc965412b8ee5e14fb220a96b4" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.121883 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.224300 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft"] Dec 01 16:14:54 crc kubenswrapper[4739]: E1201 16:14:54.224739 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f111abd9-3325-4578-8429-f107f9ba3185" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.224763 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="f111abd9-3325-4578-8429-f107f9ba3185" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.224977 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="f111abd9-3325-4578-8429-f107f9ba3185" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.225713 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.230271 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.230677 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.230932 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.231073 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft"] Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.231115 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.231155 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.231232 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.232062 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.231335 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.253124 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.253193 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.253230 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.253265 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.253321 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6zsv\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-kube-api-access-x6zsv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.253382 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.253506 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.253529 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.253562 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.253589 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.253699 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.253788 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.253830 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.355290 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.355399 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.355458 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.355508 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.355552 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.355607 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.355657 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.355723 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.355782 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.355840 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.355885 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.355936 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.356019 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6zsv\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-kube-api-access-x6zsv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.362231 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.363498 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.363689 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.365101 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.365738 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.365743 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.367655 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.367732 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.368711 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.369017 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.369078 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.369267 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.375798 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6zsv\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-kube-api-access-x6zsv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jpnft\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:54 crc kubenswrapper[4739]: I1201 16:14:54.555034 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:14:55 crc kubenswrapper[4739]: I1201 16:14:55.175058 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft"] Dec 01 16:14:55 crc kubenswrapper[4739]: W1201 16:14:55.181140 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd04ca357_3e15_4fbc_bfe1_299c3cbc2f0c.slice/crio-23bb1942af7ed6e8a19ed7becf530d91feebe945eb67f30a779b60af28b6e89d WatchSource:0}: Error finding container 23bb1942af7ed6e8a19ed7becf530d91feebe945eb67f30a779b60af28b6e89d: Status 404 returned error can't find the container with id 23bb1942af7ed6e8a19ed7becf530d91feebe945eb67f30a779b60af28b6e89d Dec 01 16:14:55 crc kubenswrapper[4739]: I1201 16:14:55.477304 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:14:55 crc kubenswrapper[4739]: E1201 16:14:55.477550 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:14:56 crc kubenswrapper[4739]: I1201 16:14:56.142345 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" event={"ID":"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c","Type":"ContainerStarted","Data":"6e2e58cd744444166c90dabbfc48605942b01b36c5f7301c2ba2f663cd387872"} Dec 01 16:14:56 crc kubenswrapper[4739]: I1201 16:14:56.143045 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" event={"ID":"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c","Type":"ContainerStarted","Data":"23bb1942af7ed6e8a19ed7becf530d91feebe945eb67f30a779b60af28b6e89d"} Dec 01 16:14:56 crc kubenswrapper[4739]: I1201 16:14:56.172524 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" podStartSLOduration=1.484122392 podStartE2EDuration="2.172503476s" podCreationTimestamp="2025-12-01 16:14:54 +0000 UTC" firstStartedPulling="2025-12-01 16:14:55.183833876 +0000 UTC m=+2397.009579990" lastFinishedPulling="2025-12-01 16:14:55.87221494 +0000 UTC m=+2397.697961074" observedRunningTime="2025-12-01 16:14:56.166468549 +0000 UTC m=+2397.992214683" watchObservedRunningTime="2025-12-01 16:14:56.172503476 +0000 UTC m=+2397.998249580" Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.149821 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn"] Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.153011 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.157815 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.158082 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.161054 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn"] Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.271715 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63ae283a-eeca-4a87-baad-8ab09d916f1e-config-volume\") pod \"collect-profiles-29410095-z4gfn\" (UID: \"63ae283a-eeca-4a87-baad-8ab09d916f1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.271787 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhbcm\" (UniqueName: \"kubernetes.io/projected/63ae283a-eeca-4a87-baad-8ab09d916f1e-kube-api-access-hhbcm\") pod \"collect-profiles-29410095-z4gfn\" (UID: \"63ae283a-eeca-4a87-baad-8ab09d916f1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.271842 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63ae283a-eeca-4a87-baad-8ab09d916f1e-secret-volume\") pod \"collect-profiles-29410095-z4gfn\" (UID: \"63ae283a-eeca-4a87-baad-8ab09d916f1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.373733 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63ae283a-eeca-4a87-baad-8ab09d916f1e-config-volume\") pod \"collect-profiles-29410095-z4gfn\" (UID: \"63ae283a-eeca-4a87-baad-8ab09d916f1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.373792 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhbcm\" (UniqueName: \"kubernetes.io/projected/63ae283a-eeca-4a87-baad-8ab09d916f1e-kube-api-access-hhbcm\") pod \"collect-profiles-29410095-z4gfn\" (UID: \"63ae283a-eeca-4a87-baad-8ab09d916f1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.373825 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63ae283a-eeca-4a87-baad-8ab09d916f1e-secret-volume\") pod \"collect-profiles-29410095-z4gfn\" (UID: \"63ae283a-eeca-4a87-baad-8ab09d916f1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.375026 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63ae283a-eeca-4a87-baad-8ab09d916f1e-config-volume\") pod \"collect-profiles-29410095-z4gfn\" (UID: \"63ae283a-eeca-4a87-baad-8ab09d916f1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.389822 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63ae283a-eeca-4a87-baad-8ab09d916f1e-secret-volume\") pod \"collect-profiles-29410095-z4gfn\" (UID: \"63ae283a-eeca-4a87-baad-8ab09d916f1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.398989 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhbcm\" (UniqueName: \"kubernetes.io/projected/63ae283a-eeca-4a87-baad-8ab09d916f1e-kube-api-access-hhbcm\") pod \"collect-profiles-29410095-z4gfn\" (UID: \"63ae283a-eeca-4a87-baad-8ab09d916f1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.481310 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" Dec 01 16:15:00 crc kubenswrapper[4739]: I1201 16:15:00.915217 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn"] Dec 01 16:15:00 crc kubenswrapper[4739]: W1201 16:15:00.923622 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63ae283a_eeca_4a87_baad_8ab09d916f1e.slice/crio-9f5c05903d1367f0bf178955f394408bf9452b27e3f4fdf261ff8439357b2048 WatchSource:0}: Error finding container 9f5c05903d1367f0bf178955f394408bf9452b27e3f4fdf261ff8439357b2048: Status 404 returned error can't find the container with id 9f5c05903d1367f0bf178955f394408bf9452b27e3f4fdf261ff8439357b2048 Dec 01 16:15:01 crc kubenswrapper[4739]: I1201 16:15:01.189945 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" event={"ID":"63ae283a-eeca-4a87-baad-8ab09d916f1e","Type":"ContainerStarted","Data":"2fe943a034e9d0086fc3b5e76fb58793cc0449ad65bc3a426e00ca6b6da6a0cc"} Dec 01 16:15:01 crc kubenswrapper[4739]: I1201 16:15:01.191308 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" event={"ID":"63ae283a-eeca-4a87-baad-8ab09d916f1e","Type":"ContainerStarted","Data":"9f5c05903d1367f0bf178955f394408bf9452b27e3f4fdf261ff8439357b2048"} Dec 01 16:15:01 crc kubenswrapper[4739]: I1201 16:15:01.208528 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" podStartSLOduration=1.208507965 podStartE2EDuration="1.208507965s" podCreationTimestamp="2025-12-01 16:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:01.204437169 +0000 UTC m=+2403.030183253" watchObservedRunningTime="2025-12-01 16:15:01.208507965 +0000 UTC m=+2403.034254059" Dec 01 16:15:02 crc kubenswrapper[4739]: I1201 16:15:02.204142 4739 generic.go:334] "Generic (PLEG): container finished" podID="63ae283a-eeca-4a87-baad-8ab09d916f1e" containerID="2fe943a034e9d0086fc3b5e76fb58793cc0449ad65bc3a426e00ca6b6da6a0cc" exitCode=0 Dec 01 16:15:02 crc kubenswrapper[4739]: I1201 16:15:02.204190 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" event={"ID":"63ae283a-eeca-4a87-baad-8ab09d916f1e","Type":"ContainerDied","Data":"2fe943a034e9d0086fc3b5e76fb58793cc0449ad65bc3a426e00ca6b6da6a0cc"} Dec 01 16:15:03 crc kubenswrapper[4739]: I1201 16:15:03.568069 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" Dec 01 16:15:03 crc kubenswrapper[4739]: I1201 16:15:03.735217 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhbcm\" (UniqueName: \"kubernetes.io/projected/63ae283a-eeca-4a87-baad-8ab09d916f1e-kube-api-access-hhbcm\") pod \"63ae283a-eeca-4a87-baad-8ab09d916f1e\" (UID: \"63ae283a-eeca-4a87-baad-8ab09d916f1e\") " Dec 01 16:15:03 crc kubenswrapper[4739]: I1201 16:15:03.735408 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63ae283a-eeca-4a87-baad-8ab09d916f1e-secret-volume\") pod \"63ae283a-eeca-4a87-baad-8ab09d916f1e\" (UID: \"63ae283a-eeca-4a87-baad-8ab09d916f1e\") " Dec 01 16:15:03 crc kubenswrapper[4739]: I1201 16:15:03.735533 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63ae283a-eeca-4a87-baad-8ab09d916f1e-config-volume\") pod \"63ae283a-eeca-4a87-baad-8ab09d916f1e\" (UID: \"63ae283a-eeca-4a87-baad-8ab09d916f1e\") " Dec 01 16:15:03 crc kubenswrapper[4739]: I1201 16:15:03.736631 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63ae283a-eeca-4a87-baad-8ab09d916f1e-config-volume" (OuterVolumeSpecName: "config-volume") pod "63ae283a-eeca-4a87-baad-8ab09d916f1e" (UID: "63ae283a-eeca-4a87-baad-8ab09d916f1e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:15:03 crc kubenswrapper[4739]: I1201 16:15:03.743817 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63ae283a-eeca-4a87-baad-8ab09d916f1e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "63ae283a-eeca-4a87-baad-8ab09d916f1e" (UID: "63ae283a-eeca-4a87-baad-8ab09d916f1e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:03 crc kubenswrapper[4739]: I1201 16:15:03.744260 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63ae283a-eeca-4a87-baad-8ab09d916f1e-kube-api-access-hhbcm" (OuterVolumeSpecName: "kube-api-access-hhbcm") pod "63ae283a-eeca-4a87-baad-8ab09d916f1e" (UID: "63ae283a-eeca-4a87-baad-8ab09d916f1e"). InnerVolumeSpecName "kube-api-access-hhbcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:03 crc kubenswrapper[4739]: I1201 16:15:03.837760 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhbcm\" (UniqueName: \"kubernetes.io/projected/63ae283a-eeca-4a87-baad-8ab09d916f1e-kube-api-access-hhbcm\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:03 crc kubenswrapper[4739]: I1201 16:15:03.837803 4739 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63ae283a-eeca-4a87-baad-8ab09d916f1e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:03 crc kubenswrapper[4739]: I1201 16:15:03.837815 4739 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63ae283a-eeca-4a87-baad-8ab09d916f1e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:04 crc kubenswrapper[4739]: I1201 16:15:04.221459 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" event={"ID":"63ae283a-eeca-4a87-baad-8ab09d916f1e","Type":"ContainerDied","Data":"9f5c05903d1367f0bf178955f394408bf9452b27e3f4fdf261ff8439357b2048"} Dec 01 16:15:04 crc kubenswrapper[4739]: I1201 16:15:04.221507 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f5c05903d1367f0bf178955f394408bf9452b27e3f4fdf261ff8439357b2048" Dec 01 16:15:04 crc kubenswrapper[4739]: I1201 16:15:04.221506 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn" Dec 01 16:15:04 crc kubenswrapper[4739]: I1201 16:15:04.276850 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv"] Dec 01 16:15:04 crc kubenswrapper[4739]: I1201 16:15:04.283267 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410050-688rv"] Dec 01 16:15:04 crc kubenswrapper[4739]: I1201 16:15:04.494280 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7e22852-525d-453f-a329-c15a1c5977ad" path="/var/lib/kubelet/pods/f7e22852-525d-453f-a329-c15a1c5977ad/volumes" Dec 01 16:15:06 crc kubenswrapper[4739]: I1201 16:15:06.478301 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:15:06 crc kubenswrapper[4739]: E1201 16:15:06.478866 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:15:18 crc kubenswrapper[4739]: I1201 16:15:18.488332 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:15:18 crc kubenswrapper[4739]: E1201 16:15:18.489575 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:15:23 crc kubenswrapper[4739]: I1201 16:15:23.790790 4739 scope.go:117] "RemoveContainer" containerID="f2a901b2ff5fe1d8f99939b2816c630f4153860853cf1793ef0f32d733a7cea1" Dec 01 16:15:29 crc kubenswrapper[4739]: I1201 16:15:29.467742 4739 generic.go:334] "Generic (PLEG): container finished" podID="d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" containerID="6e2e58cd744444166c90dabbfc48605942b01b36c5f7301c2ba2f663cd387872" exitCode=0 Dec 01 16:15:29 crc kubenswrapper[4739]: I1201 16:15:29.467779 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" event={"ID":"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c","Type":"ContainerDied","Data":"6e2e58cd744444166c90dabbfc48605942b01b36c5f7301c2ba2f663cd387872"} Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.478694 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:15:30 crc kubenswrapper[4739]: E1201 16:15:30.479061 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.869997 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.966159 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.966254 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-neutron-metadata-combined-ca-bundle\") pod \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.966315 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-repo-setup-combined-ca-bundle\") pod \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.966342 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.966368 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-libvirt-combined-ca-bundle\") pod \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.966393 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-bootstrap-combined-ca-bundle\") pod \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.966439 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ovn-combined-ca-bundle\") pod \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.966470 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-nova-combined-ca-bundle\") pod \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.966499 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-inventory\") pod \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.966537 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ssh-key\") pod \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.966622 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ceph\") pod \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.966693 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6zsv\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-kube-api-access-x6zsv\") pod \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.966722 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\" (UID: \"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c\") " Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.973288 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ceph" (OuterVolumeSpecName: "ceph") pod "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" (UID: "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.973315 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" (UID: "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.973819 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" (UID: "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.973904 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" (UID: "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.973896 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" (UID: "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.974055 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" (UID: "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.974262 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" (UID: "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.974845 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-kube-api-access-x6zsv" (OuterVolumeSpecName: "kube-api-access-x6zsv") pod "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" (UID: "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c"). InnerVolumeSpecName "kube-api-access-x6zsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.974971 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" (UID: "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.976443 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" (UID: "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.984961 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" (UID: "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.997544 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-inventory" (OuterVolumeSpecName: "inventory") pod "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" (UID: "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:30 crc kubenswrapper[4739]: I1201 16:15:30.999615 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" (UID: "d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.069826 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6zsv\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-kube-api-access-x6zsv\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.069871 4739 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.069889 4739 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.069902 4739 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.069914 4739 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.069929 4739 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.069942 4739 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.069955 4739 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.069966 4739 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.069976 4739 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.069989 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.069998 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.070008 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.490493 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" event={"ID":"d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c","Type":"ContainerDied","Data":"23bb1942af7ed6e8a19ed7becf530d91feebe945eb67f30a779b60af28b6e89d"} Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.490562 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23bb1942af7ed6e8a19ed7becf530d91feebe945eb67f30a779b60af28b6e89d" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.490568 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jpnft" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.619563 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj"] Dec 01 16:15:31 crc kubenswrapper[4739]: E1201 16:15:31.620219 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ae283a-eeca-4a87-baad-8ab09d916f1e" containerName="collect-profiles" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.620237 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ae283a-eeca-4a87-baad-8ab09d916f1e" containerName="collect-profiles" Dec 01 16:15:31 crc kubenswrapper[4739]: E1201 16:15:31.620248 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.620257 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.620492 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.620513 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ae283a-eeca-4a87-baad-8ab09d916f1e" containerName="collect-profiles" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.621191 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.624092 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.624378 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.624714 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.624746 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.624853 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.633706 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj"] Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.782157 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.782235 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.782315 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.782658 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bgsn\" (UniqueName: \"kubernetes.io/projected/7c8c89e6-a763-42d7-8e47-84d172df5822-kube-api-access-6bgsn\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.884900 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bgsn\" (UniqueName: \"kubernetes.io/projected/7c8c89e6-a763-42d7-8e47-84d172df5822-kube-api-access-6bgsn\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.885016 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.885046 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.885071 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.890612 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.892041 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.897858 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.919846 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bgsn\" (UniqueName: \"kubernetes.io/projected/7c8c89e6-a763-42d7-8e47-84d172df5822-kube-api-access-6bgsn\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:31 crc kubenswrapper[4739]: I1201 16:15:31.943640 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:32 crc kubenswrapper[4739]: I1201 16:15:32.493351 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj"] Dec 01 16:15:33 crc kubenswrapper[4739]: I1201 16:15:33.517356 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" event={"ID":"7c8c89e6-a763-42d7-8e47-84d172df5822","Type":"ContainerStarted","Data":"76385e4d6ce7648599b8226dba8a88160066c618e3060c1725e9af7161765900"} Dec 01 16:15:33 crc kubenswrapper[4739]: I1201 16:15:33.517925 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" event={"ID":"7c8c89e6-a763-42d7-8e47-84d172df5822","Type":"ContainerStarted","Data":"c44182b27488aedd6d60176291017d1112fc1ea268920ada1fa64593afd08dfa"} Dec 01 16:15:33 crc kubenswrapper[4739]: I1201 16:15:33.543161 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" podStartSLOduration=1.873898678 podStartE2EDuration="2.543101449s" podCreationTimestamp="2025-12-01 16:15:31 +0000 UTC" firstStartedPulling="2025-12-01 16:15:32.496127818 +0000 UTC m=+2434.321873922" lastFinishedPulling="2025-12-01 16:15:33.165330599 +0000 UTC m=+2434.991076693" observedRunningTime="2025-12-01 16:15:33.536929118 +0000 UTC m=+2435.362675282" watchObservedRunningTime="2025-12-01 16:15:33.543101449 +0000 UTC m=+2435.368847573" Dec 01 16:15:39 crc kubenswrapper[4739]: I1201 16:15:39.615981 4739 generic.go:334] "Generic (PLEG): container finished" podID="7c8c89e6-a763-42d7-8e47-84d172df5822" containerID="76385e4d6ce7648599b8226dba8a88160066c618e3060c1725e9af7161765900" exitCode=0 Dec 01 16:15:39 crc kubenswrapper[4739]: I1201 16:15:39.616093 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" event={"ID":"7c8c89e6-a763-42d7-8e47-84d172df5822","Type":"ContainerDied","Data":"76385e4d6ce7648599b8226dba8a88160066c618e3060c1725e9af7161765900"} Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.025831 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.170911 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-ceph\") pod \"7c8c89e6-a763-42d7-8e47-84d172df5822\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.171133 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-inventory\") pod \"7c8c89e6-a763-42d7-8e47-84d172df5822\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.171166 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-ssh-key\") pod \"7c8c89e6-a763-42d7-8e47-84d172df5822\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.171214 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bgsn\" (UniqueName: \"kubernetes.io/projected/7c8c89e6-a763-42d7-8e47-84d172df5822-kube-api-access-6bgsn\") pod \"7c8c89e6-a763-42d7-8e47-84d172df5822\" (UID: \"7c8c89e6-a763-42d7-8e47-84d172df5822\") " Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.178846 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-ceph" (OuterVolumeSpecName: "ceph") pod "7c8c89e6-a763-42d7-8e47-84d172df5822" (UID: "7c8c89e6-a763-42d7-8e47-84d172df5822"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.179323 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c8c89e6-a763-42d7-8e47-84d172df5822-kube-api-access-6bgsn" (OuterVolumeSpecName: "kube-api-access-6bgsn") pod "7c8c89e6-a763-42d7-8e47-84d172df5822" (UID: "7c8c89e6-a763-42d7-8e47-84d172df5822"). InnerVolumeSpecName "kube-api-access-6bgsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.205126 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7c8c89e6-a763-42d7-8e47-84d172df5822" (UID: "7c8c89e6-a763-42d7-8e47-84d172df5822"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.208335 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-inventory" (OuterVolumeSpecName: "inventory") pod "7c8c89e6-a763-42d7-8e47-84d172df5822" (UID: "7c8c89e6-a763-42d7-8e47-84d172df5822"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.273830 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.273864 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.273876 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bgsn\" (UniqueName: \"kubernetes.io/projected/7c8c89e6-a763-42d7-8e47-84d172df5822-kube-api-access-6bgsn\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.273888 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7c8c89e6-a763-42d7-8e47-84d172df5822-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.638025 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" event={"ID":"7c8c89e6-a763-42d7-8e47-84d172df5822","Type":"ContainerDied","Data":"c44182b27488aedd6d60176291017d1112fc1ea268920ada1fa64593afd08dfa"} Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.638410 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c44182b27488aedd6d60176291017d1112fc1ea268920ada1fa64593afd08dfa" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.638092 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.728912 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69"] Dec 01 16:15:41 crc kubenswrapper[4739]: E1201 16:15:41.729407 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c8c89e6-a763-42d7-8e47-84d172df5822" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.729452 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c8c89e6-a763-42d7-8e47-84d172df5822" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.729770 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c8c89e6-a763-42d7-8e47-84d172df5822" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.730561 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.733302 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.733359 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.733404 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.733448 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.733326 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.734930 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.738311 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69"] Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.884842 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.884927 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.884957 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.885028 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.885051 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.885268 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9frcm\" (UniqueName: \"kubernetes.io/projected/58d30160-c6ad-4f46-bc27-f7b28a07d97a-kube-api-access-9frcm\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.987280 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.987324 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.987442 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9frcm\" (UniqueName: \"kubernetes.io/projected/58d30160-c6ad-4f46-bc27-f7b28a07d97a-kube-api-access-9frcm\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.987571 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.987609 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.987633 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.988736 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.992161 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.993153 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.997118 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:41 crc kubenswrapper[4739]: I1201 16:15:41.999717 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:42 crc kubenswrapper[4739]: I1201 16:15:42.017699 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9frcm\" (UniqueName: \"kubernetes.io/projected/58d30160-c6ad-4f46-bc27-f7b28a07d97a-kube-api-access-9frcm\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5sh69\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:42 crc kubenswrapper[4739]: I1201 16:15:42.056194 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:15:42 crc kubenswrapper[4739]: I1201 16:15:42.590037 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69"] Dec 01 16:15:42 crc kubenswrapper[4739]: I1201 16:15:42.649018 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" event={"ID":"58d30160-c6ad-4f46-bc27-f7b28a07d97a","Type":"ContainerStarted","Data":"2546dfd0dad88ee62159a9dc1141dc5b6750cf3e78d1bd50fb23151d67f25ff2"} Dec 01 16:15:43 crc kubenswrapper[4739]: I1201 16:15:43.478229 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:15:43 crc kubenswrapper[4739]: E1201 16:15:43.479245 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:15:43 crc kubenswrapper[4739]: I1201 16:15:43.658708 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" event={"ID":"58d30160-c6ad-4f46-bc27-f7b28a07d97a","Type":"ContainerStarted","Data":"85a10f167352834ef82d403ee0b1c26b327cf00222416432aadcb792f7e12d75"} Dec 01 16:15:43 crc kubenswrapper[4739]: I1201 16:15:43.680125 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" podStartSLOduration=2.144303185 podStartE2EDuration="2.680098415s" podCreationTimestamp="2025-12-01 16:15:41 +0000 UTC" firstStartedPulling="2025-12-01 16:15:42.608533315 +0000 UTC m=+2444.434279399" lastFinishedPulling="2025-12-01 16:15:43.144328535 +0000 UTC m=+2444.970074629" observedRunningTime="2025-12-01 16:15:43.677102603 +0000 UTC m=+2445.502848727" watchObservedRunningTime="2025-12-01 16:15:43.680098415 +0000 UTC m=+2445.505844529" Dec 01 16:15:54 crc kubenswrapper[4739]: I1201 16:15:54.478576 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:15:54 crc kubenswrapper[4739]: E1201 16:15:54.479924 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:16:05 crc kubenswrapper[4739]: I1201 16:16:05.477136 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:16:05 crc kubenswrapper[4739]: E1201 16:16:05.478188 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:16:19 crc kubenswrapper[4739]: I1201 16:16:19.477732 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:16:19 crc kubenswrapper[4739]: I1201 16:16:19.981661 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"f6b75a6d222890688b975df140f3561a50d58b2d3f5abff948a2d8a0fc87295b"} Dec 01 16:16:59 crc kubenswrapper[4739]: I1201 16:16:59.396559 4739 generic.go:334] "Generic (PLEG): container finished" podID="58d30160-c6ad-4f46-bc27-f7b28a07d97a" containerID="85a10f167352834ef82d403ee0b1c26b327cf00222416432aadcb792f7e12d75" exitCode=0 Dec 01 16:16:59 crc kubenswrapper[4739]: I1201 16:16:59.396832 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" event={"ID":"58d30160-c6ad-4f46-bc27-f7b28a07d97a","Type":"ContainerDied","Data":"85a10f167352834ef82d403ee0b1c26b327cf00222416432aadcb792f7e12d75"} Dec 01 16:17:00 crc kubenswrapper[4739]: I1201 16:17:00.863865 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:17:00 crc kubenswrapper[4739]: I1201 16:17:00.951111 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-inventory\") pod \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " Dec 01 16:17:00 crc kubenswrapper[4739]: I1201 16:17:00.951328 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ovn-combined-ca-bundle\") pod \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " Dec 01 16:17:00 crc kubenswrapper[4739]: I1201 16:17:00.951404 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ssh-key\") pod \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " Dec 01 16:17:00 crc kubenswrapper[4739]: I1201 16:17:00.951554 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ceph\") pod \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " Dec 01 16:17:00 crc kubenswrapper[4739]: I1201 16:17:00.951618 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9frcm\" (UniqueName: \"kubernetes.io/projected/58d30160-c6ad-4f46-bc27-f7b28a07d97a-kube-api-access-9frcm\") pod \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " Dec 01 16:17:00 crc kubenswrapper[4739]: I1201 16:17:00.951669 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ovncontroller-config-0\") pod \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\" (UID: \"58d30160-c6ad-4f46-bc27-f7b28a07d97a\") " Dec 01 16:17:00 crc kubenswrapper[4739]: I1201 16:17:00.960144 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ceph" (OuterVolumeSpecName: "ceph") pod "58d30160-c6ad-4f46-bc27-f7b28a07d97a" (UID: "58d30160-c6ad-4f46-bc27-f7b28a07d97a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:17:00 crc kubenswrapper[4739]: I1201 16:17:00.960414 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "58d30160-c6ad-4f46-bc27-f7b28a07d97a" (UID: "58d30160-c6ad-4f46-bc27-f7b28a07d97a"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:17:00 crc kubenswrapper[4739]: I1201 16:17:00.960491 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58d30160-c6ad-4f46-bc27-f7b28a07d97a-kube-api-access-9frcm" (OuterVolumeSpecName: "kube-api-access-9frcm") pod "58d30160-c6ad-4f46-bc27-f7b28a07d97a" (UID: "58d30160-c6ad-4f46-bc27-f7b28a07d97a"). InnerVolumeSpecName "kube-api-access-9frcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:17:00 crc kubenswrapper[4739]: I1201 16:17:00.981960 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "58d30160-c6ad-4f46-bc27-f7b28a07d97a" (UID: "58d30160-c6ad-4f46-bc27-f7b28a07d97a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.001478 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "58d30160-c6ad-4f46-bc27-f7b28a07d97a" (UID: "58d30160-c6ad-4f46-bc27-f7b28a07d97a"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.009405 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-inventory" (OuterVolumeSpecName: "inventory") pod "58d30160-c6ad-4f46-bc27-f7b28a07d97a" (UID: "58d30160-c6ad-4f46-bc27-f7b28a07d97a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.053195 4739 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.053238 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.053250 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.053261 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9frcm\" (UniqueName: \"kubernetes.io/projected/58d30160-c6ad-4f46-bc27-f7b28a07d97a-kube-api-access-9frcm\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.053274 4739 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/58d30160-c6ad-4f46-bc27-f7b28a07d97a-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.053285 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58d30160-c6ad-4f46-bc27-f7b28a07d97a-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.423508 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" event={"ID":"58d30160-c6ad-4f46-bc27-f7b28a07d97a","Type":"ContainerDied","Data":"2546dfd0dad88ee62159a9dc1141dc5b6750cf3e78d1bd50fb23151d67f25ff2"} Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.423573 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2546dfd0dad88ee62159a9dc1141dc5b6750cf3e78d1bd50fb23151d67f25ff2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.423609 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5sh69" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.551105 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2"] Dec 01 16:17:01 crc kubenswrapper[4739]: E1201 16:17:01.551607 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d30160-c6ad-4f46-bc27-f7b28a07d97a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.551633 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d30160-c6ad-4f46-bc27-f7b28a07d97a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.551862 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d30160-c6ad-4f46-bc27-f7b28a07d97a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.552723 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.556676 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.556963 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.559951 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.560223 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.560794 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.561019 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.568439 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.571623 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2"] Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.667998 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.668400 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.668486 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.668880 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.668941 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.669169 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktvfj\" (UniqueName: \"kubernetes.io/projected/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-kube-api-access-ktvfj\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.669255 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.771263 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.771347 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.771444 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktvfj\" (UniqueName: \"kubernetes.io/projected/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-kube-api-access-ktvfj\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.771491 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.771635 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.771678 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.771718 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.777540 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.779335 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.779381 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.779681 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.782696 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.783738 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.801664 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktvfj\" (UniqueName: \"kubernetes.io/projected/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-kube-api-access-ktvfj\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:01 crc kubenswrapper[4739]: I1201 16:17:01.916503 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:17:02 crc kubenswrapper[4739]: I1201 16:17:02.276517 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2"] Dec 01 16:17:02 crc kubenswrapper[4739]: I1201 16:17:02.433581 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" event={"ID":"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4","Type":"ContainerStarted","Data":"70bcdd8f88047996b6bc81c95152fc21729d39253a470850a02087f6596ac5f3"} Dec 01 16:17:03 crc kubenswrapper[4739]: I1201 16:17:03.449304 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" event={"ID":"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4","Type":"ContainerStarted","Data":"936ff38482f2a2b9e56caa705bc4ba0e531b620a79322c95c9e58fb3cbbfebaa"} Dec 01 16:17:03 crc kubenswrapper[4739]: I1201 16:17:03.490373 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" podStartSLOduration=1.668371536 podStartE2EDuration="2.490344636s" podCreationTimestamp="2025-12-01 16:17:01 +0000 UTC" firstStartedPulling="2025-12-01 16:17:02.278641377 +0000 UTC m=+2524.104387471" lastFinishedPulling="2025-12-01 16:17:03.100614437 +0000 UTC m=+2524.926360571" observedRunningTime="2025-12-01 16:17:03.480510392 +0000 UTC m=+2525.306256536" watchObservedRunningTime="2025-12-01 16:17:03.490344636 +0000 UTC m=+2525.316090770" Dec 01 16:18:08 crc kubenswrapper[4739]: I1201 16:18:08.074866 4739 generic.go:334] "Generic (PLEG): container finished" podID="529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4" containerID="936ff38482f2a2b9e56caa705bc4ba0e531b620a79322c95c9e58fb3cbbfebaa" exitCode=0 Dec 01 16:18:08 crc kubenswrapper[4739]: I1201 16:18:08.074987 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" event={"ID":"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4","Type":"ContainerDied","Data":"936ff38482f2a2b9e56caa705bc4ba0e531b620a79322c95c9e58fb3cbbfebaa"} Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.516880 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.734591 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-inventory\") pod \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.734674 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-nova-metadata-neutron-config-0\") pod \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.734743 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-ceph\") pod \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.735107 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktvfj\" (UniqueName: \"kubernetes.io/projected/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-kube-api-access-ktvfj\") pod \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.735321 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.735390 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-neutron-metadata-combined-ca-bundle\") pod \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.736359 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-ssh-key\") pod \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\" (UID: \"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4\") " Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.741639 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-ceph" (OuterVolumeSpecName: "ceph") pod "529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4" (UID: "529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.742120 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4" (UID: "529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.743509 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-kube-api-access-ktvfj" (OuterVolumeSpecName: "kube-api-access-ktvfj") pod "529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4" (UID: "529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4"). InnerVolumeSpecName "kube-api-access-ktvfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.764599 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4" (UID: "529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.764746 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-inventory" (OuterVolumeSpecName: "inventory") pod "529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4" (UID: "529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.771463 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4" (UID: "529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.775491 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4" (UID: "529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.838317 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktvfj\" (UniqueName: \"kubernetes.io/projected/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-kube-api-access-ktvfj\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.838361 4739 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.838373 4739 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.838383 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.838392 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.838400 4739 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:09 crc kubenswrapper[4739]: I1201 16:18:09.838410 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.096892 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" event={"ID":"529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4","Type":"ContainerDied","Data":"70bcdd8f88047996b6bc81c95152fc21729d39253a470850a02087f6596ac5f3"} Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.096927 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70bcdd8f88047996b6bc81c95152fc21729d39253a470850a02087f6596ac5f3" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.096967 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.188897 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7"] Dec 01 16:18:10 crc kubenswrapper[4739]: E1201 16:18:10.189285 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.189303 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.189566 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.190154 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.192159 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.192364 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.192534 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.193778 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.196617 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.200996 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.202495 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7"] Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.345331 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.345442 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.345481 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh4gl\" (UniqueName: \"kubernetes.io/projected/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-kube-api-access-dh4gl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.345650 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.345733 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.345783 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.447287 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.447358 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.447401 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.447473 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.447495 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.447511 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh4gl\" (UniqueName: \"kubernetes.io/projected/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-kube-api-access-dh4gl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.451106 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.451106 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.451480 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.452087 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.461724 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.470146 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh4gl\" (UniqueName: \"kubernetes.io/projected/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-kube-api-access-dh4gl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-x85w7\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:10 crc kubenswrapper[4739]: I1201 16:18:10.511256 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:18:11 crc kubenswrapper[4739]: I1201 16:18:11.058658 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7"] Dec 01 16:18:11 crc kubenswrapper[4739]: I1201 16:18:11.108368 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" event={"ID":"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c","Type":"ContainerStarted","Data":"57588ef5c45356eac1707c5d0b03a0b144f2eb91659abd6c27d915f751ba0b62"} Dec 01 16:18:12 crc kubenswrapper[4739]: I1201 16:18:12.119043 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" event={"ID":"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c","Type":"ContainerStarted","Data":"3e8cc5930f5afd4907af6ea4a0483a98b5676bbbacfd1bcb0543afd3ae205175"} Dec 01 16:18:12 crc kubenswrapper[4739]: I1201 16:18:12.146639 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" podStartSLOduration=1.662022067 podStartE2EDuration="2.146597351s" podCreationTimestamp="2025-12-01 16:18:10 +0000 UTC" firstStartedPulling="2025-12-01 16:18:11.058625954 +0000 UTC m=+2592.884372068" lastFinishedPulling="2025-12-01 16:18:11.543201258 +0000 UTC m=+2593.368947352" observedRunningTime="2025-12-01 16:18:12.137287753 +0000 UTC m=+2593.963033867" watchObservedRunningTime="2025-12-01 16:18:12.146597351 +0000 UTC m=+2593.972343465" Dec 01 16:18:39 crc kubenswrapper[4739]: I1201 16:18:39.621863 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:18:39 crc kubenswrapper[4739]: I1201 16:18:39.622543 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:19:09 crc kubenswrapper[4739]: I1201 16:19:09.622198 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:19:09 crc kubenswrapper[4739]: I1201 16:19:09.622837 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:19:09 crc kubenswrapper[4739]: I1201 16:19:09.731400 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k8cdh"] Dec 01 16:19:09 crc kubenswrapper[4739]: I1201 16:19:09.733508 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:09 crc kubenswrapper[4739]: I1201 16:19:09.743012 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k8cdh"] Dec 01 16:19:09 crc kubenswrapper[4739]: I1201 16:19:09.861990 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/257c802e-44c2-4533-b0ce-0deea3be8f91-utilities\") pod \"redhat-operators-k8cdh\" (UID: \"257c802e-44c2-4533-b0ce-0deea3be8f91\") " pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:09 crc kubenswrapper[4739]: I1201 16:19:09.862848 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/257c802e-44c2-4533-b0ce-0deea3be8f91-catalog-content\") pod \"redhat-operators-k8cdh\" (UID: \"257c802e-44c2-4533-b0ce-0deea3be8f91\") " pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:09 crc kubenswrapper[4739]: I1201 16:19:09.862906 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrq4j\" (UniqueName: \"kubernetes.io/projected/257c802e-44c2-4533-b0ce-0deea3be8f91-kube-api-access-nrq4j\") pod \"redhat-operators-k8cdh\" (UID: \"257c802e-44c2-4533-b0ce-0deea3be8f91\") " pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:09 crc kubenswrapper[4739]: I1201 16:19:09.964942 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/257c802e-44c2-4533-b0ce-0deea3be8f91-utilities\") pod \"redhat-operators-k8cdh\" (UID: \"257c802e-44c2-4533-b0ce-0deea3be8f91\") " pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:09 crc kubenswrapper[4739]: I1201 16:19:09.965045 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/257c802e-44c2-4533-b0ce-0deea3be8f91-catalog-content\") pod \"redhat-operators-k8cdh\" (UID: \"257c802e-44c2-4533-b0ce-0deea3be8f91\") " pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:09 crc kubenswrapper[4739]: I1201 16:19:09.965078 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrq4j\" (UniqueName: \"kubernetes.io/projected/257c802e-44c2-4533-b0ce-0deea3be8f91-kube-api-access-nrq4j\") pod \"redhat-operators-k8cdh\" (UID: \"257c802e-44c2-4533-b0ce-0deea3be8f91\") " pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:09 crc kubenswrapper[4739]: I1201 16:19:09.965496 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/257c802e-44c2-4533-b0ce-0deea3be8f91-utilities\") pod \"redhat-operators-k8cdh\" (UID: \"257c802e-44c2-4533-b0ce-0deea3be8f91\") " pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:09 crc kubenswrapper[4739]: I1201 16:19:09.965767 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/257c802e-44c2-4533-b0ce-0deea3be8f91-catalog-content\") pod \"redhat-operators-k8cdh\" (UID: \"257c802e-44c2-4533-b0ce-0deea3be8f91\") " pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:09 crc kubenswrapper[4739]: I1201 16:19:09.996546 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrq4j\" (UniqueName: \"kubernetes.io/projected/257c802e-44c2-4533-b0ce-0deea3be8f91-kube-api-access-nrq4j\") pod \"redhat-operators-k8cdh\" (UID: \"257c802e-44c2-4533-b0ce-0deea3be8f91\") " pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:10 crc kubenswrapper[4739]: I1201 16:19:10.059805 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:10 crc kubenswrapper[4739]: I1201 16:19:10.561785 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k8cdh"] Dec 01 16:19:10 crc kubenswrapper[4739]: W1201 16:19:10.569327 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod257c802e_44c2_4533_b0ce_0deea3be8f91.slice/crio-4a05689d6d2a31d086efa41ef12321b94aabbd57ae5ff35fba761fbee778aecb WatchSource:0}: Error finding container 4a05689d6d2a31d086efa41ef12321b94aabbd57ae5ff35fba761fbee778aecb: Status 404 returned error can't find the container with id 4a05689d6d2a31d086efa41ef12321b94aabbd57ae5ff35fba761fbee778aecb Dec 01 16:19:10 crc kubenswrapper[4739]: I1201 16:19:10.677885 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8cdh" event={"ID":"257c802e-44c2-4533-b0ce-0deea3be8f91","Type":"ContainerStarted","Data":"4a05689d6d2a31d086efa41ef12321b94aabbd57ae5ff35fba761fbee778aecb"} Dec 01 16:19:11 crc kubenswrapper[4739]: I1201 16:19:11.691308 4739 generic.go:334] "Generic (PLEG): container finished" podID="257c802e-44c2-4533-b0ce-0deea3be8f91" containerID="658f9fae5cd140841bdae082c8a4a2d70735438163b3725d5b6299f4ca117bee" exitCode=0 Dec 01 16:19:11 crc kubenswrapper[4739]: I1201 16:19:11.691647 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8cdh" event={"ID":"257c802e-44c2-4533-b0ce-0deea3be8f91","Type":"ContainerDied","Data":"658f9fae5cd140841bdae082c8a4a2d70735438163b3725d5b6299f4ca117bee"} Dec 01 16:19:12 crc kubenswrapper[4739]: I1201 16:19:12.703692 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8cdh" event={"ID":"257c802e-44c2-4533-b0ce-0deea3be8f91","Type":"ContainerStarted","Data":"e8a84cf12ffb84b2c6be3d3bf77a89a339c8c040f4fcb1b93a73604fce9a0bfb"} Dec 01 16:19:14 crc kubenswrapper[4739]: I1201 16:19:14.723713 4739 generic.go:334] "Generic (PLEG): container finished" podID="257c802e-44c2-4533-b0ce-0deea3be8f91" containerID="e8a84cf12ffb84b2c6be3d3bf77a89a339c8c040f4fcb1b93a73604fce9a0bfb" exitCode=0 Dec 01 16:19:14 crc kubenswrapper[4739]: I1201 16:19:14.723926 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8cdh" event={"ID":"257c802e-44c2-4533-b0ce-0deea3be8f91","Type":"ContainerDied","Data":"e8a84cf12ffb84b2c6be3d3bf77a89a339c8c040f4fcb1b93a73604fce9a0bfb"} Dec 01 16:19:14 crc kubenswrapper[4739]: I1201 16:19:14.726923 4739 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:19:16 crc kubenswrapper[4739]: I1201 16:19:16.749985 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8cdh" event={"ID":"257c802e-44c2-4533-b0ce-0deea3be8f91","Type":"ContainerStarted","Data":"ae3d7fee4133ae13ea813c3d9f017f54000525c3d1da7235f238368f0df06d40"} Dec 01 16:19:16 crc kubenswrapper[4739]: I1201 16:19:16.777009 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k8cdh" podStartSLOduration=3.143795815 podStartE2EDuration="7.776993231s" podCreationTimestamp="2025-12-01 16:19:09 +0000 UTC" firstStartedPulling="2025-12-01 16:19:11.694539072 +0000 UTC m=+2653.520285216" lastFinishedPulling="2025-12-01 16:19:16.327736528 +0000 UTC m=+2658.153482632" observedRunningTime="2025-12-01 16:19:16.773764641 +0000 UTC m=+2658.599510745" watchObservedRunningTime="2025-12-01 16:19:16.776993231 +0000 UTC m=+2658.602739325" Dec 01 16:19:20 crc kubenswrapper[4739]: I1201 16:19:20.061708 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:20 crc kubenswrapper[4739]: I1201 16:19:20.062518 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:21 crc kubenswrapper[4739]: I1201 16:19:21.116393 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k8cdh" podUID="257c802e-44c2-4533-b0ce-0deea3be8f91" containerName="registry-server" probeResult="failure" output=< Dec 01 16:19:21 crc kubenswrapper[4739]: timeout: failed to connect service ":50051" within 1s Dec 01 16:19:21 crc kubenswrapper[4739]: > Dec 01 16:19:30 crc kubenswrapper[4739]: I1201 16:19:30.135578 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:30 crc kubenswrapper[4739]: I1201 16:19:30.210253 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:30 crc kubenswrapper[4739]: I1201 16:19:30.390675 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k8cdh"] Dec 01 16:19:31 crc kubenswrapper[4739]: I1201 16:19:31.903281 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k8cdh" podUID="257c802e-44c2-4533-b0ce-0deea3be8f91" containerName="registry-server" containerID="cri-o://ae3d7fee4133ae13ea813c3d9f017f54000525c3d1da7235f238368f0df06d40" gracePeriod=2 Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.361495 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.513218 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/257c802e-44c2-4533-b0ce-0deea3be8f91-catalog-content\") pod \"257c802e-44c2-4533-b0ce-0deea3be8f91\" (UID: \"257c802e-44c2-4533-b0ce-0deea3be8f91\") " Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.514474 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/257c802e-44c2-4533-b0ce-0deea3be8f91-utilities\") pod \"257c802e-44c2-4533-b0ce-0deea3be8f91\" (UID: \"257c802e-44c2-4533-b0ce-0deea3be8f91\") " Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.514811 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrq4j\" (UniqueName: \"kubernetes.io/projected/257c802e-44c2-4533-b0ce-0deea3be8f91-kube-api-access-nrq4j\") pod \"257c802e-44c2-4533-b0ce-0deea3be8f91\" (UID: \"257c802e-44c2-4533-b0ce-0deea3be8f91\") " Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.516820 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/257c802e-44c2-4533-b0ce-0deea3be8f91-utilities" (OuterVolumeSpecName: "utilities") pod "257c802e-44c2-4533-b0ce-0deea3be8f91" (UID: "257c802e-44c2-4533-b0ce-0deea3be8f91"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.521169 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/257c802e-44c2-4533-b0ce-0deea3be8f91-kube-api-access-nrq4j" (OuterVolumeSpecName: "kube-api-access-nrq4j") pod "257c802e-44c2-4533-b0ce-0deea3be8f91" (UID: "257c802e-44c2-4533-b0ce-0deea3be8f91"). InnerVolumeSpecName "kube-api-access-nrq4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.617598 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/257c802e-44c2-4533-b0ce-0deea3be8f91-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.617851 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrq4j\" (UniqueName: \"kubernetes.io/projected/257c802e-44c2-4533-b0ce-0deea3be8f91-kube-api-access-nrq4j\") on node \"crc\" DevicePath \"\"" Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.650979 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/257c802e-44c2-4533-b0ce-0deea3be8f91-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "257c802e-44c2-4533-b0ce-0deea3be8f91" (UID: "257c802e-44c2-4533-b0ce-0deea3be8f91"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.719899 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/257c802e-44c2-4533-b0ce-0deea3be8f91-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.917157 4739 generic.go:334] "Generic (PLEG): container finished" podID="257c802e-44c2-4533-b0ce-0deea3be8f91" containerID="ae3d7fee4133ae13ea813c3d9f017f54000525c3d1da7235f238368f0df06d40" exitCode=0 Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.917969 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k8cdh" Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.918000 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8cdh" event={"ID":"257c802e-44c2-4533-b0ce-0deea3be8f91","Type":"ContainerDied","Data":"ae3d7fee4133ae13ea813c3d9f017f54000525c3d1da7235f238368f0df06d40"} Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.919182 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k8cdh" event={"ID":"257c802e-44c2-4533-b0ce-0deea3be8f91","Type":"ContainerDied","Data":"4a05689d6d2a31d086efa41ef12321b94aabbd57ae5ff35fba761fbee778aecb"} Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.919219 4739 scope.go:117] "RemoveContainer" containerID="ae3d7fee4133ae13ea813c3d9f017f54000525c3d1da7235f238368f0df06d40" Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.951830 4739 scope.go:117] "RemoveContainer" containerID="e8a84cf12ffb84b2c6be3d3bf77a89a339c8c040f4fcb1b93a73604fce9a0bfb" Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.974197 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k8cdh"] Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.987245 4739 scope.go:117] "RemoveContainer" containerID="658f9fae5cd140841bdae082c8a4a2d70735438163b3725d5b6299f4ca117bee" Dec 01 16:19:32 crc kubenswrapper[4739]: I1201 16:19:32.988493 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k8cdh"] Dec 01 16:19:33 crc kubenswrapper[4739]: I1201 16:19:33.022776 4739 scope.go:117] "RemoveContainer" containerID="ae3d7fee4133ae13ea813c3d9f017f54000525c3d1da7235f238368f0df06d40" Dec 01 16:19:33 crc kubenswrapper[4739]: E1201 16:19:33.023342 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae3d7fee4133ae13ea813c3d9f017f54000525c3d1da7235f238368f0df06d40\": container with ID starting with ae3d7fee4133ae13ea813c3d9f017f54000525c3d1da7235f238368f0df06d40 not found: ID does not exist" containerID="ae3d7fee4133ae13ea813c3d9f017f54000525c3d1da7235f238368f0df06d40" Dec 01 16:19:33 crc kubenswrapper[4739]: I1201 16:19:33.023379 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae3d7fee4133ae13ea813c3d9f017f54000525c3d1da7235f238368f0df06d40"} err="failed to get container status \"ae3d7fee4133ae13ea813c3d9f017f54000525c3d1da7235f238368f0df06d40\": rpc error: code = NotFound desc = could not find container \"ae3d7fee4133ae13ea813c3d9f017f54000525c3d1da7235f238368f0df06d40\": container with ID starting with ae3d7fee4133ae13ea813c3d9f017f54000525c3d1da7235f238368f0df06d40 not found: ID does not exist" Dec 01 16:19:33 crc kubenswrapper[4739]: I1201 16:19:33.023407 4739 scope.go:117] "RemoveContainer" containerID="e8a84cf12ffb84b2c6be3d3bf77a89a339c8c040f4fcb1b93a73604fce9a0bfb" Dec 01 16:19:33 crc kubenswrapper[4739]: E1201 16:19:33.024759 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8a84cf12ffb84b2c6be3d3bf77a89a339c8c040f4fcb1b93a73604fce9a0bfb\": container with ID starting with e8a84cf12ffb84b2c6be3d3bf77a89a339c8c040f4fcb1b93a73604fce9a0bfb not found: ID does not exist" containerID="e8a84cf12ffb84b2c6be3d3bf77a89a339c8c040f4fcb1b93a73604fce9a0bfb" Dec 01 16:19:33 crc kubenswrapper[4739]: I1201 16:19:33.024795 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8a84cf12ffb84b2c6be3d3bf77a89a339c8c040f4fcb1b93a73604fce9a0bfb"} err="failed to get container status \"e8a84cf12ffb84b2c6be3d3bf77a89a339c8c040f4fcb1b93a73604fce9a0bfb\": rpc error: code = NotFound desc = could not find container \"e8a84cf12ffb84b2c6be3d3bf77a89a339c8c040f4fcb1b93a73604fce9a0bfb\": container with ID starting with e8a84cf12ffb84b2c6be3d3bf77a89a339c8c040f4fcb1b93a73604fce9a0bfb not found: ID does not exist" Dec 01 16:19:33 crc kubenswrapper[4739]: I1201 16:19:33.024813 4739 scope.go:117] "RemoveContainer" containerID="658f9fae5cd140841bdae082c8a4a2d70735438163b3725d5b6299f4ca117bee" Dec 01 16:19:33 crc kubenswrapper[4739]: E1201 16:19:33.025297 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"658f9fae5cd140841bdae082c8a4a2d70735438163b3725d5b6299f4ca117bee\": container with ID starting with 658f9fae5cd140841bdae082c8a4a2d70735438163b3725d5b6299f4ca117bee not found: ID does not exist" containerID="658f9fae5cd140841bdae082c8a4a2d70735438163b3725d5b6299f4ca117bee" Dec 01 16:19:33 crc kubenswrapper[4739]: I1201 16:19:33.025324 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"658f9fae5cd140841bdae082c8a4a2d70735438163b3725d5b6299f4ca117bee"} err="failed to get container status \"658f9fae5cd140841bdae082c8a4a2d70735438163b3725d5b6299f4ca117bee\": rpc error: code = NotFound desc = could not find container \"658f9fae5cd140841bdae082c8a4a2d70735438163b3725d5b6299f4ca117bee\": container with ID starting with 658f9fae5cd140841bdae082c8a4a2d70735438163b3725d5b6299f4ca117bee not found: ID does not exist" Dec 01 16:19:34 crc kubenswrapper[4739]: I1201 16:19:34.501025 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="257c802e-44c2-4533-b0ce-0deea3be8f91" path="/var/lib/kubelet/pods/257c802e-44c2-4533-b0ce-0deea3be8f91/volumes" Dec 01 16:19:39 crc kubenswrapper[4739]: I1201 16:19:39.622223 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:19:39 crc kubenswrapper[4739]: I1201 16:19:39.622991 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:19:39 crc kubenswrapper[4739]: I1201 16:19:39.623044 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 16:19:39 crc kubenswrapper[4739]: I1201 16:19:39.623892 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f6b75a6d222890688b975df140f3561a50d58b2d3f5abff948a2d8a0fc87295b"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:19:39 crc kubenswrapper[4739]: I1201 16:19:39.623963 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://f6b75a6d222890688b975df140f3561a50d58b2d3f5abff948a2d8a0fc87295b" gracePeriod=600 Dec 01 16:19:40 crc kubenswrapper[4739]: I1201 16:19:40.005374 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="f6b75a6d222890688b975df140f3561a50d58b2d3f5abff948a2d8a0fc87295b" exitCode=0 Dec 01 16:19:40 crc kubenswrapper[4739]: I1201 16:19:40.005468 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"f6b75a6d222890688b975df140f3561a50d58b2d3f5abff948a2d8a0fc87295b"} Dec 01 16:19:40 crc kubenswrapper[4739]: I1201 16:19:40.005803 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d"} Dec 01 16:19:40 crc kubenswrapper[4739]: I1201 16:19:40.005832 4739 scope.go:117] "RemoveContainer" containerID="2816292afa2f8d77d038ecf222dce1c9ecfbc767fa4281d4adc05102739316be" Dec 01 16:19:51 crc kubenswrapper[4739]: I1201 16:19:51.288753 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-t98c7" podUID="19abbd38-da74-4e94-94c5-9461701ac88d" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 16:21:39 crc kubenswrapper[4739]: I1201 16:21:39.622106 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:21:39 crc kubenswrapper[4739]: I1201 16:21:39.623715 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.286165 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-98rtn"] Dec 01 16:21:59 crc kubenswrapper[4739]: E1201 16:21:59.287227 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="257c802e-44c2-4533-b0ce-0deea3be8f91" containerName="extract-utilities" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.287245 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="257c802e-44c2-4533-b0ce-0deea3be8f91" containerName="extract-utilities" Dec 01 16:21:59 crc kubenswrapper[4739]: E1201 16:21:59.287271 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="257c802e-44c2-4533-b0ce-0deea3be8f91" containerName="extract-content" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.287279 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="257c802e-44c2-4533-b0ce-0deea3be8f91" containerName="extract-content" Dec 01 16:21:59 crc kubenswrapper[4739]: E1201 16:21:59.287304 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="257c802e-44c2-4533-b0ce-0deea3be8f91" containerName="registry-server" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.287314 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="257c802e-44c2-4533-b0ce-0deea3be8f91" containerName="registry-server" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.287547 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="257c802e-44c2-4533-b0ce-0deea3be8f91" containerName="registry-server" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.289104 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.303273 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-98rtn"] Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.409703 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85qnj\" (UniqueName: \"kubernetes.io/projected/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-kube-api-access-85qnj\") pod \"redhat-marketplace-98rtn\" (UID: \"c4ac15dc-c679-43fe-bafb-88bf102ec9f3\") " pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.409763 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-utilities\") pod \"redhat-marketplace-98rtn\" (UID: \"c4ac15dc-c679-43fe-bafb-88bf102ec9f3\") " pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.409875 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-catalog-content\") pod \"redhat-marketplace-98rtn\" (UID: \"c4ac15dc-c679-43fe-bafb-88bf102ec9f3\") " pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.511819 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-utilities\") pod \"redhat-marketplace-98rtn\" (UID: \"c4ac15dc-c679-43fe-bafb-88bf102ec9f3\") " pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.512000 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-catalog-content\") pod \"redhat-marketplace-98rtn\" (UID: \"c4ac15dc-c679-43fe-bafb-88bf102ec9f3\") " pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.512231 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85qnj\" (UniqueName: \"kubernetes.io/projected/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-kube-api-access-85qnj\") pod \"redhat-marketplace-98rtn\" (UID: \"c4ac15dc-c679-43fe-bafb-88bf102ec9f3\") " pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.512743 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-utilities\") pod \"redhat-marketplace-98rtn\" (UID: \"c4ac15dc-c679-43fe-bafb-88bf102ec9f3\") " pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.512830 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-catalog-content\") pod \"redhat-marketplace-98rtn\" (UID: \"c4ac15dc-c679-43fe-bafb-88bf102ec9f3\") " pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.534518 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85qnj\" (UniqueName: \"kubernetes.io/projected/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-kube-api-access-85qnj\") pod \"redhat-marketplace-98rtn\" (UID: \"c4ac15dc-c679-43fe-bafb-88bf102ec9f3\") " pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:21:59 crc kubenswrapper[4739]: I1201 16:21:59.616889 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:22:00 crc kubenswrapper[4739]: I1201 16:22:00.138513 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-98rtn"] Dec 01 16:22:00 crc kubenswrapper[4739]: I1201 16:22:00.414140 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98rtn" event={"ID":"c4ac15dc-c679-43fe-bafb-88bf102ec9f3","Type":"ContainerStarted","Data":"936b2f1e60219ac718acff0abda96fc24694b57e65c9f4c0a86930cffbc81c5d"} Dec 01 16:22:00 crc kubenswrapper[4739]: I1201 16:22:00.414602 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98rtn" event={"ID":"c4ac15dc-c679-43fe-bafb-88bf102ec9f3","Type":"ContainerStarted","Data":"1a84090d9a3d7fbe558821465e01dee211f0359e932f1c79f0dafd4dcb84a153"} Dec 01 16:22:01 crc kubenswrapper[4739]: I1201 16:22:01.429291 4739 generic.go:334] "Generic (PLEG): container finished" podID="c4ac15dc-c679-43fe-bafb-88bf102ec9f3" containerID="936b2f1e60219ac718acff0abda96fc24694b57e65c9f4c0a86930cffbc81c5d" exitCode=0 Dec 01 16:22:01 crc kubenswrapper[4739]: I1201 16:22:01.429692 4739 generic.go:334] "Generic (PLEG): container finished" podID="c4ac15dc-c679-43fe-bafb-88bf102ec9f3" containerID="0e4dfd2b48f3b0db28f9f68b4ef12dc10a58896c375f6107259f2fbdf2a35400" exitCode=0 Dec 01 16:22:01 crc kubenswrapper[4739]: I1201 16:22:01.429394 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98rtn" event={"ID":"c4ac15dc-c679-43fe-bafb-88bf102ec9f3","Type":"ContainerDied","Data":"936b2f1e60219ac718acff0abda96fc24694b57e65c9f4c0a86930cffbc81c5d"} Dec 01 16:22:01 crc kubenswrapper[4739]: I1201 16:22:01.429751 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98rtn" event={"ID":"c4ac15dc-c679-43fe-bafb-88bf102ec9f3","Type":"ContainerDied","Data":"0e4dfd2b48f3b0db28f9f68b4ef12dc10a58896c375f6107259f2fbdf2a35400"} Dec 01 16:22:02 crc kubenswrapper[4739]: I1201 16:22:02.438758 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98rtn" event={"ID":"c4ac15dc-c679-43fe-bafb-88bf102ec9f3","Type":"ContainerStarted","Data":"094762cbbb44ec805030f172cdb9d6be27733a15d628f72c47e359ec32339ce4"} Dec 01 16:22:02 crc kubenswrapper[4739]: I1201 16:22:02.459738 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-98rtn" podStartSLOduration=2.047515589 podStartE2EDuration="3.459718229s" podCreationTimestamp="2025-12-01 16:21:59 +0000 UTC" firstStartedPulling="2025-12-01 16:22:00.417667909 +0000 UTC m=+2822.243414053" lastFinishedPulling="2025-12-01 16:22:01.829870559 +0000 UTC m=+2823.655616693" observedRunningTime="2025-12-01 16:22:02.45490453 +0000 UTC m=+2824.280650624" watchObservedRunningTime="2025-12-01 16:22:02.459718229 +0000 UTC m=+2824.285464323" Dec 01 16:22:07 crc kubenswrapper[4739]: I1201 16:22:07.968751 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t42x6"] Dec 01 16:22:07 crc kubenswrapper[4739]: I1201 16:22:07.971715 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:07 crc kubenswrapper[4739]: I1201 16:22:07.982704 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t42x6"] Dec 01 16:22:08 crc kubenswrapper[4739]: I1201 16:22:08.081276 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl8pp\" (UniqueName: \"kubernetes.io/projected/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-kube-api-access-sl8pp\") pod \"certified-operators-t42x6\" (UID: \"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892\") " pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:08 crc kubenswrapper[4739]: I1201 16:22:08.081326 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-utilities\") pod \"certified-operators-t42x6\" (UID: \"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892\") " pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:08 crc kubenswrapper[4739]: I1201 16:22:08.081437 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-catalog-content\") pod \"certified-operators-t42x6\" (UID: \"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892\") " pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:08 crc kubenswrapper[4739]: I1201 16:22:08.182781 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl8pp\" (UniqueName: \"kubernetes.io/projected/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-kube-api-access-sl8pp\") pod \"certified-operators-t42x6\" (UID: \"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892\") " pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:08 crc kubenswrapper[4739]: I1201 16:22:08.182843 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-utilities\") pod \"certified-operators-t42x6\" (UID: \"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892\") " pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:08 crc kubenswrapper[4739]: I1201 16:22:08.182890 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-catalog-content\") pod \"certified-operators-t42x6\" (UID: \"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892\") " pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:08 crc kubenswrapper[4739]: I1201 16:22:08.183458 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-catalog-content\") pod \"certified-operators-t42x6\" (UID: \"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892\") " pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:08 crc kubenswrapper[4739]: I1201 16:22:08.183979 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-utilities\") pod \"certified-operators-t42x6\" (UID: \"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892\") " pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:08 crc kubenswrapper[4739]: I1201 16:22:08.203073 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl8pp\" (UniqueName: \"kubernetes.io/projected/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-kube-api-access-sl8pp\") pod \"certified-operators-t42x6\" (UID: \"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892\") " pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:08 crc kubenswrapper[4739]: I1201 16:22:08.304557 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:08 crc kubenswrapper[4739]: I1201 16:22:08.822029 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t42x6"] Dec 01 16:22:09 crc kubenswrapper[4739]: I1201 16:22:09.522753 4739 generic.go:334] "Generic (PLEG): container finished" podID="1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892" containerID="bec9eef7e986008ba08d1ae2dd26124ddcb2fa48fe824a3c669dee34c643bfc8" exitCode=0 Dec 01 16:22:09 crc kubenswrapper[4739]: I1201 16:22:09.523028 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t42x6" event={"ID":"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892","Type":"ContainerDied","Data":"bec9eef7e986008ba08d1ae2dd26124ddcb2fa48fe824a3c669dee34c643bfc8"} Dec 01 16:22:09 crc kubenswrapper[4739]: I1201 16:22:09.523176 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t42x6" event={"ID":"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892","Type":"ContainerStarted","Data":"1f4e3d3abff8a583e1dd9ec3bb328b486c8e3ac90d30da64163cb4626ff32953"} Dec 01 16:22:09 crc kubenswrapper[4739]: I1201 16:22:09.617672 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:22:09 crc kubenswrapper[4739]: I1201 16:22:09.617738 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:22:09 crc kubenswrapper[4739]: I1201 16:22:09.622256 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:22:09 crc kubenswrapper[4739]: I1201 16:22:09.622309 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:22:09 crc kubenswrapper[4739]: I1201 16:22:09.694346 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:22:10 crc kubenswrapper[4739]: I1201 16:22:10.535543 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t42x6" event={"ID":"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892","Type":"ContainerStarted","Data":"36009a05a9f9bf20656d6e5c1b3c5a4d5994cc4d7d8db9086a7fd1061e78c8e6"} Dec 01 16:22:10 crc kubenswrapper[4739]: I1201 16:22:10.622480 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:22:11 crc kubenswrapper[4739]: I1201 16:22:11.551361 4739 generic.go:334] "Generic (PLEG): container finished" podID="1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892" containerID="36009a05a9f9bf20656d6e5c1b3c5a4d5994cc4d7d8db9086a7fd1061e78c8e6" exitCode=0 Dec 01 16:22:11 crc kubenswrapper[4739]: I1201 16:22:11.551585 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t42x6" event={"ID":"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892","Type":"ContainerDied","Data":"36009a05a9f9bf20656d6e5c1b3c5a4d5994cc4d7d8db9086a7fd1061e78c8e6"} Dec 01 16:22:11 crc kubenswrapper[4739]: I1201 16:22:11.953939 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-98rtn"] Dec 01 16:22:12 crc kubenswrapper[4739]: I1201 16:22:12.563066 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t42x6" event={"ID":"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892","Type":"ContainerStarted","Data":"8202623f57048612272e3f410ff6db60d6a6edd1e794ab8eb1d6cfe57af7a881"} Dec 01 16:22:12 crc kubenswrapper[4739]: I1201 16:22:12.563206 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-98rtn" podUID="c4ac15dc-c679-43fe-bafb-88bf102ec9f3" containerName="registry-server" containerID="cri-o://094762cbbb44ec805030f172cdb9d6be27733a15d628f72c47e359ec32339ce4" gracePeriod=2 Dec 01 16:22:12 crc kubenswrapper[4739]: I1201 16:22:12.597769 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t42x6" podStartSLOduration=2.97940419 podStartE2EDuration="5.597748047s" podCreationTimestamp="2025-12-01 16:22:07 +0000 UTC" firstStartedPulling="2025-12-01 16:22:09.525199015 +0000 UTC m=+2831.350945109" lastFinishedPulling="2025-12-01 16:22:12.143542842 +0000 UTC m=+2833.969288966" observedRunningTime="2025-12-01 16:22:12.587051137 +0000 UTC m=+2834.412797241" watchObservedRunningTime="2025-12-01 16:22:12.597748047 +0000 UTC m=+2834.423494151" Dec 01 16:22:12 crc kubenswrapper[4739]: I1201 16:22:12.997817 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.075388 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-catalog-content\") pod \"c4ac15dc-c679-43fe-bafb-88bf102ec9f3\" (UID: \"c4ac15dc-c679-43fe-bafb-88bf102ec9f3\") " Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.075569 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85qnj\" (UniqueName: \"kubernetes.io/projected/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-kube-api-access-85qnj\") pod \"c4ac15dc-c679-43fe-bafb-88bf102ec9f3\" (UID: \"c4ac15dc-c679-43fe-bafb-88bf102ec9f3\") " Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.075656 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-utilities\") pod \"c4ac15dc-c679-43fe-bafb-88bf102ec9f3\" (UID: \"c4ac15dc-c679-43fe-bafb-88bf102ec9f3\") " Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.076720 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-utilities" (OuterVolumeSpecName: "utilities") pod "c4ac15dc-c679-43fe-bafb-88bf102ec9f3" (UID: "c4ac15dc-c679-43fe-bafb-88bf102ec9f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.081944 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-kube-api-access-85qnj" (OuterVolumeSpecName: "kube-api-access-85qnj") pod "c4ac15dc-c679-43fe-bafb-88bf102ec9f3" (UID: "c4ac15dc-c679-43fe-bafb-88bf102ec9f3"). InnerVolumeSpecName "kube-api-access-85qnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.102455 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4ac15dc-c679-43fe-bafb-88bf102ec9f3" (UID: "c4ac15dc-c679-43fe-bafb-88bf102ec9f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.178251 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.178381 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.178406 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85qnj\" (UniqueName: \"kubernetes.io/projected/c4ac15dc-c679-43fe-bafb-88bf102ec9f3-kube-api-access-85qnj\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.575474 4739 generic.go:334] "Generic (PLEG): container finished" podID="c4ac15dc-c679-43fe-bafb-88bf102ec9f3" containerID="094762cbbb44ec805030f172cdb9d6be27733a15d628f72c47e359ec32339ce4" exitCode=0 Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.575572 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98rtn" event={"ID":"c4ac15dc-c679-43fe-bafb-88bf102ec9f3","Type":"ContainerDied","Data":"094762cbbb44ec805030f172cdb9d6be27733a15d628f72c47e359ec32339ce4"} Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.575617 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98rtn" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.575666 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98rtn" event={"ID":"c4ac15dc-c679-43fe-bafb-88bf102ec9f3","Type":"ContainerDied","Data":"1a84090d9a3d7fbe558821465e01dee211f0359e932f1c79f0dafd4dcb84a153"} Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.575692 4739 scope.go:117] "RemoveContainer" containerID="094762cbbb44ec805030f172cdb9d6be27733a15d628f72c47e359ec32339ce4" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.615875 4739 scope.go:117] "RemoveContainer" containerID="0e4dfd2b48f3b0db28f9f68b4ef12dc10a58896c375f6107259f2fbdf2a35400" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.622815 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-98rtn"] Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.631921 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-98rtn"] Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.663751 4739 scope.go:117] "RemoveContainer" containerID="936b2f1e60219ac718acff0abda96fc24694b57e65c9f4c0a86930cffbc81c5d" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.701656 4739 scope.go:117] "RemoveContainer" containerID="094762cbbb44ec805030f172cdb9d6be27733a15d628f72c47e359ec32339ce4" Dec 01 16:22:13 crc kubenswrapper[4739]: E1201 16:22:13.702152 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"094762cbbb44ec805030f172cdb9d6be27733a15d628f72c47e359ec32339ce4\": container with ID starting with 094762cbbb44ec805030f172cdb9d6be27733a15d628f72c47e359ec32339ce4 not found: ID does not exist" containerID="094762cbbb44ec805030f172cdb9d6be27733a15d628f72c47e359ec32339ce4" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.702303 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"094762cbbb44ec805030f172cdb9d6be27733a15d628f72c47e359ec32339ce4"} err="failed to get container status \"094762cbbb44ec805030f172cdb9d6be27733a15d628f72c47e359ec32339ce4\": rpc error: code = NotFound desc = could not find container \"094762cbbb44ec805030f172cdb9d6be27733a15d628f72c47e359ec32339ce4\": container with ID starting with 094762cbbb44ec805030f172cdb9d6be27733a15d628f72c47e359ec32339ce4 not found: ID does not exist" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.702354 4739 scope.go:117] "RemoveContainer" containerID="0e4dfd2b48f3b0db28f9f68b4ef12dc10a58896c375f6107259f2fbdf2a35400" Dec 01 16:22:13 crc kubenswrapper[4739]: E1201 16:22:13.702739 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e4dfd2b48f3b0db28f9f68b4ef12dc10a58896c375f6107259f2fbdf2a35400\": container with ID starting with 0e4dfd2b48f3b0db28f9f68b4ef12dc10a58896c375f6107259f2fbdf2a35400 not found: ID does not exist" containerID="0e4dfd2b48f3b0db28f9f68b4ef12dc10a58896c375f6107259f2fbdf2a35400" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.702844 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e4dfd2b48f3b0db28f9f68b4ef12dc10a58896c375f6107259f2fbdf2a35400"} err="failed to get container status \"0e4dfd2b48f3b0db28f9f68b4ef12dc10a58896c375f6107259f2fbdf2a35400\": rpc error: code = NotFound desc = could not find container \"0e4dfd2b48f3b0db28f9f68b4ef12dc10a58896c375f6107259f2fbdf2a35400\": container with ID starting with 0e4dfd2b48f3b0db28f9f68b4ef12dc10a58896c375f6107259f2fbdf2a35400 not found: ID does not exist" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.702881 4739 scope.go:117] "RemoveContainer" containerID="936b2f1e60219ac718acff0abda96fc24694b57e65c9f4c0a86930cffbc81c5d" Dec 01 16:22:13 crc kubenswrapper[4739]: E1201 16:22:13.703183 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"936b2f1e60219ac718acff0abda96fc24694b57e65c9f4c0a86930cffbc81c5d\": container with ID starting with 936b2f1e60219ac718acff0abda96fc24694b57e65c9f4c0a86930cffbc81c5d not found: ID does not exist" containerID="936b2f1e60219ac718acff0abda96fc24694b57e65c9f4c0a86930cffbc81c5d" Dec 01 16:22:13 crc kubenswrapper[4739]: I1201 16:22:13.703217 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"936b2f1e60219ac718acff0abda96fc24694b57e65c9f4c0a86930cffbc81c5d"} err="failed to get container status \"936b2f1e60219ac718acff0abda96fc24694b57e65c9f4c0a86930cffbc81c5d\": rpc error: code = NotFound desc = could not find container \"936b2f1e60219ac718acff0abda96fc24694b57e65c9f4c0a86930cffbc81c5d\": container with ID starting with 936b2f1e60219ac718acff0abda96fc24694b57e65c9f4c0a86930cffbc81c5d not found: ID does not exist" Dec 01 16:22:14 crc kubenswrapper[4739]: I1201 16:22:14.494686 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4ac15dc-c679-43fe-bafb-88bf102ec9f3" path="/var/lib/kubelet/pods/c4ac15dc-c679-43fe-bafb-88bf102ec9f3/volumes" Dec 01 16:22:18 crc kubenswrapper[4739]: I1201 16:22:18.306350 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:18 crc kubenswrapper[4739]: I1201 16:22:18.307065 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:18 crc kubenswrapper[4739]: I1201 16:22:18.361166 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:18 crc kubenswrapper[4739]: I1201 16:22:18.696923 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:18 crc kubenswrapper[4739]: I1201 16:22:18.748374 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t42x6"] Dec 01 16:22:20 crc kubenswrapper[4739]: I1201 16:22:20.646346 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t42x6" podUID="1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892" containerName="registry-server" containerID="cri-o://8202623f57048612272e3f410ff6db60d6a6edd1e794ab8eb1d6cfe57af7a881" gracePeriod=2 Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.119508 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.234357 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-utilities\") pod \"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892\" (UID: \"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892\") " Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.235050 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-catalog-content\") pod \"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892\" (UID: \"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892\") " Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.235219 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sl8pp\" (UniqueName: \"kubernetes.io/projected/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-kube-api-access-sl8pp\") pod \"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892\" (UID: \"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892\") " Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.236009 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-utilities" (OuterVolumeSpecName: "utilities") pod "1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892" (UID: "1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.246179 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-kube-api-access-sl8pp" (OuterVolumeSpecName: "kube-api-access-sl8pp") pod "1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892" (UID: "1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892"). InnerVolumeSpecName "kube-api-access-sl8pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.337473 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sl8pp\" (UniqueName: \"kubernetes.io/projected/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-kube-api-access-sl8pp\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.337507 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.449793 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892" (UID: "1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.541579 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.659412 4739 generic.go:334] "Generic (PLEG): container finished" podID="1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892" containerID="8202623f57048612272e3f410ff6db60d6a6edd1e794ab8eb1d6cfe57af7a881" exitCode=0 Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.659497 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t42x6" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.659499 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t42x6" event={"ID":"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892","Type":"ContainerDied","Data":"8202623f57048612272e3f410ff6db60d6a6edd1e794ab8eb1d6cfe57af7a881"} Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.659544 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t42x6" event={"ID":"1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892","Type":"ContainerDied","Data":"1f4e3d3abff8a583e1dd9ec3bb328b486c8e3ac90d30da64163cb4626ff32953"} Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.659574 4739 scope.go:117] "RemoveContainer" containerID="8202623f57048612272e3f410ff6db60d6a6edd1e794ab8eb1d6cfe57af7a881" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.685063 4739 scope.go:117] "RemoveContainer" containerID="36009a05a9f9bf20656d6e5c1b3c5a4d5994cc4d7d8db9086a7fd1061e78c8e6" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.700119 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t42x6"] Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.708118 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t42x6"] Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.730611 4739 scope.go:117] "RemoveContainer" containerID="bec9eef7e986008ba08d1ae2dd26124ddcb2fa48fe824a3c669dee34c643bfc8" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.751012 4739 scope.go:117] "RemoveContainer" containerID="8202623f57048612272e3f410ff6db60d6a6edd1e794ab8eb1d6cfe57af7a881" Dec 01 16:22:21 crc kubenswrapper[4739]: E1201 16:22:21.751649 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8202623f57048612272e3f410ff6db60d6a6edd1e794ab8eb1d6cfe57af7a881\": container with ID starting with 8202623f57048612272e3f410ff6db60d6a6edd1e794ab8eb1d6cfe57af7a881 not found: ID does not exist" containerID="8202623f57048612272e3f410ff6db60d6a6edd1e794ab8eb1d6cfe57af7a881" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.751692 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8202623f57048612272e3f410ff6db60d6a6edd1e794ab8eb1d6cfe57af7a881"} err="failed to get container status \"8202623f57048612272e3f410ff6db60d6a6edd1e794ab8eb1d6cfe57af7a881\": rpc error: code = NotFound desc = could not find container \"8202623f57048612272e3f410ff6db60d6a6edd1e794ab8eb1d6cfe57af7a881\": container with ID starting with 8202623f57048612272e3f410ff6db60d6a6edd1e794ab8eb1d6cfe57af7a881 not found: ID does not exist" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.751726 4739 scope.go:117] "RemoveContainer" containerID="36009a05a9f9bf20656d6e5c1b3c5a4d5994cc4d7d8db9086a7fd1061e78c8e6" Dec 01 16:22:21 crc kubenswrapper[4739]: E1201 16:22:21.752059 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36009a05a9f9bf20656d6e5c1b3c5a4d5994cc4d7d8db9086a7fd1061e78c8e6\": container with ID starting with 36009a05a9f9bf20656d6e5c1b3c5a4d5994cc4d7d8db9086a7fd1061e78c8e6 not found: ID does not exist" containerID="36009a05a9f9bf20656d6e5c1b3c5a4d5994cc4d7d8db9086a7fd1061e78c8e6" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.752088 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36009a05a9f9bf20656d6e5c1b3c5a4d5994cc4d7d8db9086a7fd1061e78c8e6"} err="failed to get container status \"36009a05a9f9bf20656d6e5c1b3c5a4d5994cc4d7d8db9086a7fd1061e78c8e6\": rpc error: code = NotFound desc = could not find container \"36009a05a9f9bf20656d6e5c1b3c5a4d5994cc4d7d8db9086a7fd1061e78c8e6\": container with ID starting with 36009a05a9f9bf20656d6e5c1b3c5a4d5994cc4d7d8db9086a7fd1061e78c8e6 not found: ID does not exist" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.752107 4739 scope.go:117] "RemoveContainer" containerID="bec9eef7e986008ba08d1ae2dd26124ddcb2fa48fe824a3c669dee34c643bfc8" Dec 01 16:22:21 crc kubenswrapper[4739]: E1201 16:22:21.752412 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bec9eef7e986008ba08d1ae2dd26124ddcb2fa48fe824a3c669dee34c643bfc8\": container with ID starting with bec9eef7e986008ba08d1ae2dd26124ddcb2fa48fe824a3c669dee34c643bfc8 not found: ID does not exist" containerID="bec9eef7e986008ba08d1ae2dd26124ddcb2fa48fe824a3c669dee34c643bfc8" Dec 01 16:22:21 crc kubenswrapper[4739]: I1201 16:22:21.752466 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bec9eef7e986008ba08d1ae2dd26124ddcb2fa48fe824a3c669dee34c643bfc8"} err="failed to get container status \"bec9eef7e986008ba08d1ae2dd26124ddcb2fa48fe824a3c669dee34c643bfc8\": rpc error: code = NotFound desc = could not find container \"bec9eef7e986008ba08d1ae2dd26124ddcb2fa48fe824a3c669dee34c643bfc8\": container with ID starting with bec9eef7e986008ba08d1ae2dd26124ddcb2fa48fe824a3c669dee34c643bfc8 not found: ID does not exist" Dec 01 16:22:22 crc kubenswrapper[4739]: I1201 16:22:22.495248 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892" path="/var/lib/kubelet/pods/1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892/volumes" Dec 01 16:22:39 crc kubenswrapper[4739]: I1201 16:22:39.621973 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:22:39 crc kubenswrapper[4739]: I1201 16:22:39.622789 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:22:39 crc kubenswrapper[4739]: I1201 16:22:39.622860 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 16:22:39 crc kubenswrapper[4739]: I1201 16:22:39.623901 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:22:39 crc kubenswrapper[4739]: I1201 16:22:39.623984 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" gracePeriod=600 Dec 01 16:22:39 crc kubenswrapper[4739]: E1201 16:22:39.747784 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:22:39 crc kubenswrapper[4739]: I1201 16:22:39.869741 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" exitCode=0 Dec 01 16:22:39 crc kubenswrapper[4739]: I1201 16:22:39.869799 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d"} Dec 01 16:22:39 crc kubenswrapper[4739]: I1201 16:22:39.869852 4739 scope.go:117] "RemoveContainer" containerID="f6b75a6d222890688b975df140f3561a50d58b2d3f5abff948a2d8a0fc87295b" Dec 01 16:22:39 crc kubenswrapper[4739]: I1201 16:22:39.870778 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:22:39 crc kubenswrapper[4739]: E1201 16:22:39.871255 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:22:49 crc kubenswrapper[4739]: I1201 16:22:49.967339 4739 generic.go:334] "Generic (PLEG): container finished" podID="cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c" containerID="3e8cc5930f5afd4907af6ea4a0483a98b5676bbbacfd1bcb0543afd3ae205175" exitCode=0 Dec 01 16:22:49 crc kubenswrapper[4739]: I1201 16:22:49.967432 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" event={"ID":"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c","Type":"ContainerDied","Data":"3e8cc5930f5afd4907af6ea4a0483a98b5676bbbacfd1bcb0543afd3ae205175"} Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.444858 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.491122 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-inventory\") pod \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.491311 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dh4gl\" (UniqueName: \"kubernetes.io/projected/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-kube-api-access-dh4gl\") pod \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.491345 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-libvirt-secret-0\") pod \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.491406 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-ssh-key\") pod \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.491457 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-libvirt-combined-ca-bundle\") pod \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.491491 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-ceph\") pod \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\" (UID: \"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c\") " Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.497375 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-kube-api-access-dh4gl" (OuterVolumeSpecName: "kube-api-access-dh4gl") pod "cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c" (UID: "cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c"). InnerVolumeSpecName "kube-api-access-dh4gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.497521 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-ceph" (OuterVolumeSpecName: "ceph") pod "cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c" (UID: "cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.498355 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c" (UID: "cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.518335 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c" (UID: "cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.518656 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c" (UID: "cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.529674 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-inventory" (OuterVolumeSpecName: "inventory") pod "cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c" (UID: "cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.593985 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dh4gl\" (UniqueName: \"kubernetes.io/projected/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-kube-api-access-dh4gl\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.594020 4739 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.594031 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.594043 4739 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.594054 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.594063 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.997731 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" event={"ID":"cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c","Type":"ContainerDied","Data":"57588ef5c45356eac1707c5d0b03a0b144f2eb91659abd6c27d915f751ba0b62"} Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.997779 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57588ef5c45356eac1707c5d0b03a0b144f2eb91659abd6c27d915f751ba0b62" Dec 01 16:22:51 crc kubenswrapper[4739]: I1201 16:22:51.997813 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-x85w7" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.120299 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4"] Dec 01 16:22:52 crc kubenswrapper[4739]: E1201 16:22:52.120811 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ac15dc-c679-43fe-bafb-88bf102ec9f3" containerName="registry-server" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.120833 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ac15dc-c679-43fe-bafb-88bf102ec9f3" containerName="registry-server" Dec 01 16:22:52 crc kubenswrapper[4739]: E1201 16:22:52.120849 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ac15dc-c679-43fe-bafb-88bf102ec9f3" containerName="extract-utilities" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.120857 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ac15dc-c679-43fe-bafb-88bf102ec9f3" containerName="extract-utilities" Dec 01 16:22:52 crc kubenswrapper[4739]: E1201 16:22:52.120876 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.120885 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 16:22:52 crc kubenswrapper[4739]: E1201 16:22:52.120906 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892" containerName="extract-content" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.120912 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892" containerName="extract-content" Dec 01 16:22:52 crc kubenswrapper[4739]: E1201 16:22:52.120927 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892" containerName="registry-server" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.120935 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892" containerName="registry-server" Dec 01 16:22:52 crc kubenswrapper[4739]: E1201 16:22:52.120950 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ac15dc-c679-43fe-bafb-88bf102ec9f3" containerName="extract-content" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.120957 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ac15dc-c679-43fe-bafb-88bf102ec9f3" containerName="extract-content" Dec 01 16:22:52 crc kubenswrapper[4739]: E1201 16:22:52.120967 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892" containerName="extract-utilities" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.120973 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892" containerName="extract-utilities" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.121169 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.121190 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ac15dc-c679-43fe-bafb-88bf102ec9f3" containerName="registry-server" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.121206 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea07ec4-c7c1-4bb6-a743-ab6e8ef9e892" containerName="registry-server" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.121932 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.127712 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.128231 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.128469 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.128638 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.129024 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qdj2c" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.129189 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.129295 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.129503 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.129835 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.134068 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4"] Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.207727 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.208110 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.208133 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.208168 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/ad6c9132-a47c-486c-bb34-17d977bef91b-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.208189 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m77ct\" (UniqueName: \"kubernetes.io/projected/ad6c9132-a47c-486c-bb34-17d977bef91b-kube-api-access-m77ct\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.208208 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.208239 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.208403 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.208477 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.208553 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.208750 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.310611 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/ad6c9132-a47c-486c-bb34-17d977bef91b-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.310661 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m77ct\" (UniqueName: \"kubernetes.io/projected/ad6c9132-a47c-486c-bb34-17d977bef91b-kube-api-access-m77ct\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.310690 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.310735 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.310774 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.310801 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.310836 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.310886 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.310957 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.310990 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.311011 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.312196 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/ad6c9132-a47c-486c-bb34-17d977bef91b-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.312304 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.314689 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.315103 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.315314 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.315886 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.316794 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.316989 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.319336 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.321275 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.328543 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m77ct\" (UniqueName: \"kubernetes.io/projected/ad6c9132-a47c-486c-bb34-17d977bef91b-kube-api-access-m77ct\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:52 crc kubenswrapper[4739]: I1201 16:22:52.445496 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:22:53 crc kubenswrapper[4739]: I1201 16:22:53.037381 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4"] Dec 01 16:22:53 crc kubenswrapper[4739]: I1201 16:22:53.478282 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:22:53 crc kubenswrapper[4739]: E1201 16:22:53.478975 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:22:54 crc kubenswrapper[4739]: I1201 16:22:54.033230 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" event={"ID":"ad6c9132-a47c-486c-bb34-17d977bef91b","Type":"ContainerStarted","Data":"8a5473ffac3c4fc3598e70b70cd131384ff5ab374f36d60057f746ed06e20c8c"} Dec 01 16:22:54 crc kubenswrapper[4739]: I1201 16:22:54.033758 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" event={"ID":"ad6c9132-a47c-486c-bb34-17d977bef91b","Type":"ContainerStarted","Data":"656003e519879481f285452a2bb9a3829c31152b4a24d6bd2913da1a732ced52"} Dec 01 16:22:54 crc kubenswrapper[4739]: I1201 16:22:54.080075 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" podStartSLOduration=1.571164234 podStartE2EDuration="2.08004402s" podCreationTimestamp="2025-12-01 16:22:52 +0000 UTC" firstStartedPulling="2025-12-01 16:22:53.045256923 +0000 UTC m=+2874.871003027" lastFinishedPulling="2025-12-01 16:22:53.554136719 +0000 UTC m=+2875.379882813" observedRunningTime="2025-12-01 16:22:54.069126363 +0000 UTC m=+2875.894872467" watchObservedRunningTime="2025-12-01 16:22:54.08004402 +0000 UTC m=+2875.905790124" Dec 01 16:23:04 crc kubenswrapper[4739]: I1201 16:23:04.477348 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:23:04 crc kubenswrapper[4739]: E1201 16:23:04.479067 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:23:17 crc kubenswrapper[4739]: I1201 16:23:17.477107 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:23:17 crc kubenswrapper[4739]: E1201 16:23:17.478345 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:23:31 crc kubenswrapper[4739]: I1201 16:23:31.477593 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:23:31 crc kubenswrapper[4739]: E1201 16:23:31.478627 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:23:43 crc kubenswrapper[4739]: I1201 16:23:43.478100 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:23:43 crc kubenswrapper[4739]: E1201 16:23:43.479169 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:23:58 crc kubenswrapper[4739]: I1201 16:23:58.494905 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:23:58 crc kubenswrapper[4739]: E1201 16:23:58.496008 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:24:12 crc kubenswrapper[4739]: I1201 16:24:12.477394 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:24:12 crc kubenswrapper[4739]: E1201 16:24:12.478398 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:24:26 crc kubenswrapper[4739]: I1201 16:24:26.477522 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:24:26 crc kubenswrapper[4739]: E1201 16:24:26.478868 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:24:37 crc kubenswrapper[4739]: I1201 16:24:37.477091 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:24:37 crc kubenswrapper[4739]: E1201 16:24:37.478985 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:24:52 crc kubenswrapper[4739]: I1201 16:24:52.477869 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:24:52 crc kubenswrapper[4739]: E1201 16:24:52.478788 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:25:07 crc kubenswrapper[4739]: I1201 16:25:07.478708 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:25:07 crc kubenswrapper[4739]: E1201 16:25:07.479870 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:25:22 crc kubenswrapper[4739]: I1201 16:25:22.478277 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:25:22 crc kubenswrapper[4739]: E1201 16:25:22.478999 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:25:36 crc kubenswrapper[4739]: I1201 16:25:36.477853 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:25:36 crc kubenswrapper[4739]: E1201 16:25:36.478992 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:25:48 crc kubenswrapper[4739]: I1201 16:25:48.482893 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:25:48 crc kubenswrapper[4739]: E1201 16:25:48.483678 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:25:59 crc kubenswrapper[4739]: I1201 16:25:59.478108 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:25:59 crc kubenswrapper[4739]: E1201 16:25:59.479205 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:26:09 crc kubenswrapper[4739]: I1201 16:26:09.033800 4739 generic.go:334] "Generic (PLEG): container finished" podID="ad6c9132-a47c-486c-bb34-17d977bef91b" containerID="8a5473ffac3c4fc3598e70b70cd131384ff5ab374f36d60057f746ed06e20c8c" exitCode=0 Dec 01 16:26:09 crc kubenswrapper[4739]: I1201 16:26:09.033916 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" event={"ID":"ad6c9132-a47c-486c-bb34-17d977bef91b","Type":"ContainerDied","Data":"8a5473ffac3c4fc3598e70b70cd131384ff5ab374f36d60057f746ed06e20c8c"} Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.413747 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.450483 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-cell1-compute-config-1\") pod \"ad6c9132-a47c-486c-bb34-17d977bef91b\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.450584 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m77ct\" (UniqueName: \"kubernetes.io/projected/ad6c9132-a47c-486c-bb34-17d977bef91b-kube-api-access-m77ct\") pod \"ad6c9132-a47c-486c-bb34-17d977bef91b\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.450617 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/ad6c9132-a47c-486c-bb34-17d977bef91b-ceph-nova-0\") pod \"ad6c9132-a47c-486c-bb34-17d977bef91b\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.450649 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-migration-ssh-key-0\") pod \"ad6c9132-a47c-486c-bb34-17d977bef91b\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.450685 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-ceph\") pod \"ad6c9132-a47c-486c-bb34-17d977bef91b\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.451398 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-cell1-compute-config-0\") pod \"ad6c9132-a47c-486c-bb34-17d977bef91b\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.451439 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-migration-ssh-key-1\") pod \"ad6c9132-a47c-486c-bb34-17d977bef91b\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.451460 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-extra-config-0\") pod \"ad6c9132-a47c-486c-bb34-17d977bef91b\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.451558 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-ssh-key\") pod \"ad6c9132-a47c-486c-bb34-17d977bef91b\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.451598 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-inventory\") pod \"ad6c9132-a47c-486c-bb34-17d977bef91b\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.451662 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-custom-ceph-combined-ca-bundle\") pod \"ad6c9132-a47c-486c-bb34-17d977bef91b\" (UID: \"ad6c9132-a47c-486c-bb34-17d977bef91b\") " Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.456128 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-ceph" (OuterVolumeSpecName: "ceph") pod "ad6c9132-a47c-486c-bb34-17d977bef91b" (UID: "ad6c9132-a47c-486c-bb34-17d977bef91b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.470051 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad6c9132-a47c-486c-bb34-17d977bef91b-kube-api-access-m77ct" (OuterVolumeSpecName: "kube-api-access-m77ct") pod "ad6c9132-a47c-486c-bb34-17d977bef91b" (UID: "ad6c9132-a47c-486c-bb34-17d977bef91b"). InnerVolumeSpecName "kube-api-access-m77ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.479555 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "ad6c9132-a47c-486c-bb34-17d977bef91b" (UID: "ad6c9132-a47c-486c-bb34-17d977bef91b"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.479939 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "ad6c9132-a47c-486c-bb34-17d977bef91b" (UID: "ad6c9132-a47c-486c-bb34-17d977bef91b"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.481688 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "ad6c9132-a47c-486c-bb34-17d977bef91b" (UID: "ad6c9132-a47c-486c-bb34-17d977bef91b"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.482046 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad6c9132-a47c-486c-bb34-17d977bef91b-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "ad6c9132-a47c-486c-bb34-17d977bef91b" (UID: "ad6c9132-a47c-486c-bb34-17d977bef91b"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.484072 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "ad6c9132-a47c-486c-bb34-17d977bef91b" (UID: "ad6c9132-a47c-486c-bb34-17d977bef91b"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.510050 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ad6c9132-a47c-486c-bb34-17d977bef91b" (UID: "ad6c9132-a47c-486c-bb34-17d977bef91b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.511755 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-inventory" (OuterVolumeSpecName: "inventory") pod "ad6c9132-a47c-486c-bb34-17d977bef91b" (UID: "ad6c9132-a47c-486c-bb34-17d977bef91b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.513813 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "ad6c9132-a47c-486c-bb34-17d977bef91b" (UID: "ad6c9132-a47c-486c-bb34-17d977bef91b"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.515865 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "ad6c9132-a47c-486c-bb34-17d977bef91b" (UID: "ad6c9132-a47c-486c-bb34-17d977bef91b"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.554025 4739 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.554242 4739 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.554385 4739 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.554481 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.554550 4739 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.554628 4739 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.554694 4739 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.554752 4739 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/ad6c9132-a47c-486c-bb34-17d977bef91b-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.554810 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m77ct\" (UniqueName: \"kubernetes.io/projected/ad6c9132-a47c-486c-bb34-17d977bef91b-kube-api-access-m77ct\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.554874 4739 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:10 crc kubenswrapper[4739]: I1201 16:26:10.554928 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ad6c9132-a47c-486c-bb34-17d977bef91b-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:11 crc kubenswrapper[4739]: I1201 16:26:11.052985 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" event={"ID":"ad6c9132-a47c-486c-bb34-17d977bef91b","Type":"ContainerDied","Data":"656003e519879481f285452a2bb9a3829c31152b4a24d6bd2913da1a732ced52"} Dec 01 16:26:11 crc kubenswrapper[4739]: I1201 16:26:11.053025 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="656003e519879481f285452a2bb9a3829c31152b4a24d6bd2913da1a732ced52" Dec 01 16:26:11 crc kubenswrapper[4739]: I1201 16:26:11.053031 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4" Dec 01 16:26:11 crc kubenswrapper[4739]: E1201 16:26:11.161735 4739 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad6c9132_a47c_486c_bb34_17d977bef91b.slice\": RecentStats: unable to find data in memory cache]" Dec 01 16:26:12 crc kubenswrapper[4739]: I1201 16:26:12.478111 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:26:12 crc kubenswrapper[4739]: E1201 16:26:12.478951 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.300046 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 01 16:26:25 crc kubenswrapper[4739]: E1201 16:26:25.301327 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad6c9132-a47c-486c-bb34-17d977bef91b" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.301356 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad6c9132-a47c-486c-bb34-17d977bef91b" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.301689 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad6c9132-a47c-486c-bb34-17d977bef91b" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.303284 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.305519 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.305746 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.311439 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.313484 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.317049 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.328709 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.341620 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.465742 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.465808 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/694f34a3-1680-4a0f-9320-9c0d83d957a7-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.465842 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/694f34a3-1680-4a0f-9320-9c0d83d957a7-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.465978 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466033 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466066 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466099 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/694f34a3-1680-4a0f-9320-9c0d83d957a7-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466250 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466294 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466364 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5r5h\" (UniqueName: \"kubernetes.io/projected/0630ef7a-359c-49a5-9906-b7c9e01936f7-kube-api-access-w5r5h\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466446 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-dev\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466498 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0630ef7a-359c-49a5-9906-b7c9e01936f7-config-data\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466597 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-lib-modules\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466624 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466691 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0630ef7a-359c-49a5-9906-b7c9e01936f7-ceph\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466744 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/694f34a3-1680-4a0f-9320-9c0d83d957a7-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466775 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w8qv\" (UniqueName: \"kubernetes.io/projected/694f34a3-1680-4a0f-9320-9c0d83d957a7-kube-api-access-8w8qv\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466887 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-run\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466942 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-sys\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.466963 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.467000 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0630ef7a-359c-49a5-9906-b7c9e01936f7-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.467023 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.467039 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.467122 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-run\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.467141 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-sys\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.467158 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/694f34a3-1680-4a0f-9320-9c0d83d957a7-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.467177 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0630ef7a-359c-49a5-9906-b7c9e01936f7-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.467199 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0630ef7a-359c-49a5-9906-b7c9e01936f7-scripts\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.467253 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.467309 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.467334 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-dev\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.467386 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.569864 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0630ef7a-359c-49a5-9906-b7c9e01936f7-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570025 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0630ef7a-359c-49a5-9906-b7c9e01936f7-scripts\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570054 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570089 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570127 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-dev\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570164 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570203 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570242 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/694f34a3-1680-4a0f-9320-9c0d83d957a7-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570286 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/694f34a3-1680-4a0f-9320-9c0d83d957a7-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570322 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570331 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570354 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570444 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570475 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/694f34a3-1680-4a0f-9320-9c0d83d957a7-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570520 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570546 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570617 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5r5h\" (UniqueName: \"kubernetes.io/projected/0630ef7a-359c-49a5-9906-b7c9e01936f7-kube-api-access-w5r5h\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570631 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570677 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-dev\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570705 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-dev\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570723 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0630ef7a-359c-49a5-9906-b7c9e01936f7-config-data\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570742 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570779 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570801 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-lib-modules\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570822 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570867 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0630ef7a-359c-49a5-9906-b7c9e01936f7-ceph\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570917 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/694f34a3-1680-4a0f-9320-9c0d83d957a7-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570935 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w8qv\" (UniqueName: \"kubernetes.io/projected/694f34a3-1680-4a0f-9320-9c0d83d957a7-kube-api-access-8w8qv\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.570970 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-run\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.571022 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-sys\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.571035 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.571049 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.571238 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0630ef7a-359c-49a5-9906-b7c9e01936f7-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.571272 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.571301 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.571327 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.571356 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.571513 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-run\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.571554 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-sys\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.571580 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/694f34a3-1680-4a0f-9320-9c0d83d957a7-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.574070 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-lib-modules\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.574080 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.574332 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.575581 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-dev\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.575630 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.575877 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-run\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.575907 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.576943 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0630ef7a-359c-49a5-9906-b7c9e01936f7-scripts\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.576999 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.577040 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/694f34a3-1680-4a0f-9320-9c0d83d957a7-sys\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.577067 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-run\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.577099 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.577131 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0630ef7a-359c-49a5-9906-b7c9e01936f7-sys\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.579879 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/694f34a3-1680-4a0f-9320-9c0d83d957a7-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.581101 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/694f34a3-1680-4a0f-9320-9c0d83d957a7-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.581227 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/694f34a3-1680-4a0f-9320-9c0d83d957a7-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.581256 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0630ef7a-359c-49a5-9906-b7c9e01936f7-ceph\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.581740 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/694f34a3-1680-4a0f-9320-9c0d83d957a7-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.582947 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0630ef7a-359c-49a5-9906-b7c9e01936f7-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.584405 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/694f34a3-1680-4a0f-9320-9c0d83d957a7-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.586870 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0630ef7a-359c-49a5-9906-b7c9e01936f7-config-data\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.591051 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0630ef7a-359c-49a5-9906-b7c9e01936f7-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.592910 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5r5h\" (UniqueName: \"kubernetes.io/projected/0630ef7a-359c-49a5-9906-b7c9e01936f7-kube-api-access-w5r5h\") pod \"cinder-backup-0\" (UID: \"0630ef7a-359c-49a5-9906-b7c9e01936f7\") " pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.598180 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w8qv\" (UniqueName: \"kubernetes.io/projected/694f34a3-1680-4a0f-9320-9c0d83d957a7-kube-api-access-8w8qv\") pod \"cinder-volume-volume1-0\" (UID: \"694f34a3-1680-4a0f-9320-9c0d83d957a7\") " pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.628995 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.653765 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.850561 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-28srx"] Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.851984 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-28srx" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.857039 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-28srx"] Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.882800 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9a7eb0a-9824-40c2-b44f-2ce4deb7604e-operator-scripts\") pod \"manila-db-create-28srx\" (UID: \"d9a7eb0a-9824-40c2-b44f-2ce4deb7604e\") " pod="openstack/manila-db-create-28srx" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.883148 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q42ql\" (UniqueName: \"kubernetes.io/projected/d9a7eb0a-9824-40c2-b44f-2ce4deb7604e-kube-api-access-q42ql\") pod \"manila-db-create-28srx\" (UID: \"d9a7eb0a-9824-40c2-b44f-2ce4deb7604e\") " pod="openstack/manila-db-create-28srx" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.905524 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-29c8-account-create-update-nd9g7"] Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.906573 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-29c8-account-create-update-nd9g7" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.909260 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.926583 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-29c8-account-create-update-nd9g7"] Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.948481 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f84667477-69xkr"] Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.950140 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.953588 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-9tm8x" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.953917 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.954046 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.954169 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.958940 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f84667477-69xkr"] Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.985227 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9a7eb0a-9824-40c2-b44f-2ce4deb7604e-operator-scripts\") pod \"manila-db-create-28srx\" (UID: \"d9a7eb0a-9824-40c2-b44f-2ce4deb7604e\") " pod="openstack/manila-db-create-28srx" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.985279 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvwxn\" (UniqueName: \"kubernetes.io/projected/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-kube-api-access-qvwxn\") pod \"horizon-7f84667477-69xkr\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.985299 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-config-data\") pod \"horizon-7f84667477-69xkr\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.985317 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-scripts\") pod \"horizon-7f84667477-69xkr\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.985357 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jtzb\" (UniqueName: \"kubernetes.io/projected/321be3c7-476a-4d82-8bde-4a5d54f5ab9d-kube-api-access-2jtzb\") pod \"manila-29c8-account-create-update-nd9g7\" (UID: \"321be3c7-476a-4d82-8bde-4a5d54f5ab9d\") " pod="openstack/manila-29c8-account-create-update-nd9g7" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.985474 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q42ql\" (UniqueName: \"kubernetes.io/projected/d9a7eb0a-9824-40c2-b44f-2ce4deb7604e-kube-api-access-q42ql\") pod \"manila-db-create-28srx\" (UID: \"d9a7eb0a-9824-40c2-b44f-2ce4deb7604e\") " pod="openstack/manila-db-create-28srx" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.985530 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-logs\") pod \"horizon-7f84667477-69xkr\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.985641 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-horizon-secret-key\") pod \"horizon-7f84667477-69xkr\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.985685 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/321be3c7-476a-4d82-8bde-4a5d54f5ab9d-operator-scripts\") pod \"manila-29c8-account-create-update-nd9g7\" (UID: \"321be3c7-476a-4d82-8bde-4a5d54f5ab9d\") " pod="openstack/manila-29c8-account-create-update-nd9g7" Dec 01 16:26:25 crc kubenswrapper[4739]: I1201 16:26:25.986177 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9a7eb0a-9824-40c2-b44f-2ce4deb7604e-operator-scripts\") pod \"manila-db-create-28srx\" (UID: \"d9a7eb0a-9824-40c2-b44f-2ce4deb7604e\") " pod="openstack/manila-db-create-28srx" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.025177 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q42ql\" (UniqueName: \"kubernetes.io/projected/d9a7eb0a-9824-40c2-b44f-2ce4deb7604e-kube-api-access-q42ql\") pod \"manila-db-create-28srx\" (UID: \"d9a7eb0a-9824-40c2-b44f-2ce4deb7604e\") " pod="openstack/manila-db-create-28srx" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.045458 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.066949 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.067241 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.070778 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.071006 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-src6f" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.071167 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.074377 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.082241 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f698fbd6f-nmnlm"] Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.083647 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.088043 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jtzb\" (UniqueName: \"kubernetes.io/projected/321be3c7-476a-4d82-8bde-4a5d54f5ab9d-kube-api-access-2jtzb\") pod \"manila-29c8-account-create-update-nd9g7\" (UID: \"321be3c7-476a-4d82-8bde-4a5d54f5ab9d\") " pod="openstack/manila-29c8-account-create-update-nd9g7" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.088085 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e9e319ab-2c20-4c46-910d-980a0221a139-horizon-secret-key\") pod \"horizon-7f698fbd6f-nmnlm\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.088118 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9e319ab-2c20-4c46-910d-980a0221a139-scripts\") pod \"horizon-7f698fbd6f-nmnlm\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.088142 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9e319ab-2c20-4c46-910d-980a0221a139-logs\") pod \"horizon-7f698fbd6f-nmnlm\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.088159 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-logs\") pod \"horizon-7f84667477-69xkr\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.088198 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z85v9\" (UniqueName: \"kubernetes.io/projected/e9e319ab-2c20-4c46-910d-980a0221a139-kube-api-access-z85v9\") pod \"horizon-7f698fbd6f-nmnlm\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.088224 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-horizon-secret-key\") pod \"horizon-7f84667477-69xkr\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.088247 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e9e319ab-2c20-4c46-910d-980a0221a139-config-data\") pod \"horizon-7f698fbd6f-nmnlm\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.088269 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/321be3c7-476a-4d82-8bde-4a5d54f5ab9d-operator-scripts\") pod \"manila-29c8-account-create-update-nd9g7\" (UID: \"321be3c7-476a-4d82-8bde-4a5d54f5ab9d\") " pod="openstack/manila-29c8-account-create-update-nd9g7" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.088320 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvwxn\" (UniqueName: \"kubernetes.io/projected/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-kube-api-access-qvwxn\") pod \"horizon-7f84667477-69xkr\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.088340 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-config-data\") pod \"horizon-7f84667477-69xkr\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.088356 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-scripts\") pod \"horizon-7f84667477-69xkr\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.090623 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-logs\") pod \"horizon-7f84667477-69xkr\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.091131 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/321be3c7-476a-4d82-8bde-4a5d54f5ab9d-operator-scripts\") pod \"manila-29c8-account-create-update-nd9g7\" (UID: \"321be3c7-476a-4d82-8bde-4a5d54f5ab9d\") " pod="openstack/manila-29c8-account-create-update-nd9g7" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.091838 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-scripts\") pod \"horizon-7f84667477-69xkr\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.093731 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-config-data\") pod \"horizon-7f84667477-69xkr\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.108235 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-horizon-secret-key\") pod \"horizon-7f84667477-69xkr\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.134466 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jtzb\" (UniqueName: \"kubernetes.io/projected/321be3c7-476a-4d82-8bde-4a5d54f5ab9d-kube-api-access-2jtzb\") pod \"manila-29c8-account-create-update-nd9g7\" (UID: \"321be3c7-476a-4d82-8bde-4a5d54f5ab9d\") " pod="openstack/manila-29c8-account-create-update-nd9g7" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.142345 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.144703 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.145531 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvwxn\" (UniqueName: \"kubernetes.io/projected/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-kube-api-access-qvwxn\") pod \"horizon-7f84667477-69xkr\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.153839 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.154101 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.156867 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f698fbd6f-nmnlm"] Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.183908 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.208275 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-28srx" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.213769 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-config-data\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.213846 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-scripts\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.213871 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.213988 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e9e319ab-2c20-4c46-910d-980a0221a139-horizon-secret-key\") pod \"horizon-7f698fbd6f-nmnlm\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.214080 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-ceph\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.214201 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9e319ab-2c20-4c46-910d-980a0221a139-scripts\") pod \"horizon-7f698fbd6f-nmnlm\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.215610 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.215769 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9e319ab-2c20-4c46-910d-980a0221a139-logs\") pod \"horizon-7f698fbd6f-nmnlm\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.215929 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-logs\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.216451 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z85v9\" (UniqueName: \"kubernetes.io/projected/e9e319ab-2c20-4c46-910d-980a0221a139-kube-api-access-z85v9\") pod \"horizon-7f698fbd6f-nmnlm\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.216635 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.216783 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e9e319ab-2c20-4c46-910d-980a0221a139-config-data\") pod \"horizon-7f698fbd6f-nmnlm\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.216931 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bw2g\" (UniqueName: \"kubernetes.io/projected/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-kube-api-access-6bw2g\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.217404 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.219403 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9e319ab-2c20-4c46-910d-980a0221a139-scripts\") pod \"horizon-7f698fbd6f-nmnlm\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.248129 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e9e319ab-2c20-4c46-910d-980a0221a139-horizon-secret-key\") pod \"horizon-7f698fbd6f-nmnlm\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.249330 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-29c8-account-create-update-nd9g7" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.249992 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9e319ab-2c20-4c46-910d-980a0221a139-logs\") pod \"horizon-7f698fbd6f-nmnlm\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.250508 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e9e319ab-2c20-4c46-910d-980a0221a139-config-data\") pod \"horizon-7f698fbd6f-nmnlm\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.268202 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.270762 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z85v9\" (UniqueName: \"kubernetes.io/projected/e9e319ab-2c20-4c46-910d-980a0221a139-kube-api-access-z85v9\") pod \"horizon-7f698fbd6f-nmnlm\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.304914 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322393 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bw2g\" (UniqueName: \"kubernetes.io/projected/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-kube-api-access-6bw2g\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322476 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322513 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322543 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322564 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bd36ab4-d8ce-4902-99d6-eb5865a70030-logs\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322591 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bd36ab4-d8ce-4902-99d6-eb5865a70030-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322616 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322635 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-config-data\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322654 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7bd36ab4-d8ce-4902-99d6-eb5865a70030-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322671 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-scripts\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322691 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322721 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-ceph\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322756 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322779 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c4tq\" (UniqueName: \"kubernetes.io/projected/7bd36ab4-d8ce-4902-99d6-eb5865a70030-kube-api-access-4c4tq\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322799 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-logs\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322820 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322845 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.322882 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.324041 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-logs\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.324488 4739 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.328352 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-config-data\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.329090 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.329313 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.329528 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.331337 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-ceph\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.332202 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-scripts\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.335145 4739 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.343603 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bw2g\" (UniqueName: \"kubernetes.io/projected/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-kube-api-access-6bw2g\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.395213 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.425589 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.425649 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.425687 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.425707 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bd36ab4-d8ce-4902-99d6-eb5865a70030-logs\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.425741 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bd36ab4-d8ce-4902-99d6-eb5865a70030-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.425784 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7bd36ab4-d8ce-4902-99d6-eb5865a70030-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.425855 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c4tq\" (UniqueName: \"kubernetes.io/projected/7bd36ab4-d8ce-4902-99d6-eb5865a70030-kube-api-access-4c4tq\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.425885 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.425918 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.427893 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bd36ab4-d8ce-4902-99d6-eb5865a70030-logs\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.428352 4739 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.431630 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.432561 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.432801 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bd36ab4-d8ce-4902-99d6-eb5865a70030-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.433344 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.435225 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7bd36ab4-d8ce-4902-99d6-eb5865a70030-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.439347 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.450177 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c4tq\" (UniqueName: \"kubernetes.io/projected/7bd36ab4-d8ce-4902-99d6-eb5865a70030-kube-api-access-4c4tq\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.456028 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.495504 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.697223 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.742618 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-28srx"] Dec 01 16:26:26 crc kubenswrapper[4739]: W1201 16:26:26.746518 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9a7eb0a_9824_40c2_b44f_2ce4deb7604e.slice/crio-1670f778a378787a3743031cefa77a928c315a821c4490d9eee5de400f06402a WatchSource:0}: Error finding container 1670f778a378787a3743031cefa77a928c315a821c4490d9eee5de400f06402a: Status 404 returned error can't find the container with id 1670f778a378787a3743031cefa77a928c315a821c4490d9eee5de400f06402a Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.798027 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.856101 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-29c8-account-create-update-nd9g7"] Dec 01 16:26:26 crc kubenswrapper[4739]: W1201 16:26:26.859594 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod321be3c7_476a_4d82_8bde_4a5d54f5ab9d.slice/crio-514e5421c9fa642d3936ec8cd32884109dedf1e35fbc19ea38a099de1375ced0 WatchSource:0}: Error finding container 514e5421c9fa642d3936ec8cd32884109dedf1e35fbc19ea38a099de1375ced0: Status 404 returned error can't find the container with id 514e5421c9fa642d3936ec8cd32884109dedf1e35fbc19ea38a099de1375ced0 Dec 01 16:26:26 crc kubenswrapper[4739]: I1201 16:26:26.959612 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f84667477-69xkr"] Dec 01 16:26:27 crc kubenswrapper[4739]: I1201 16:26:27.010293 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f698fbd6f-nmnlm"] Dec 01 16:26:27 crc kubenswrapper[4739]: W1201 16:26:27.022966 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9e319ab_2c20_4c46_910d_980a0221a139.slice/crio-c93ca7e3a8547c86666650f9320fe20376fc31c4992f7b624240ba00fe24ae9c WatchSource:0}: Error finding container c93ca7e3a8547c86666650f9320fe20376fc31c4992f7b624240ba00fe24ae9c: Status 404 returned error can't find the container with id c93ca7e3a8547c86666650f9320fe20376fc31c4992f7b624240ba00fe24ae9c Dec 01 16:26:27 crc kubenswrapper[4739]: I1201 16:26:27.221523 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 01 16:26:27 crc kubenswrapper[4739]: W1201 16:26:27.226752 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod694f34a3_1680_4a0f_9320_9c0d83d957a7.slice/crio-cda85e24a449975ce71458cbb42fd5bfe9b3ae4ac903ecce9da34694e2d7b561 WatchSource:0}: Error finding container cda85e24a449975ce71458cbb42fd5bfe9b3ae4ac903ecce9da34694e2d7b561: Status 404 returned error can't find the container with id cda85e24a449975ce71458cbb42fd5bfe9b3ae4ac903ecce9da34694e2d7b561 Dec 01 16:26:27 crc kubenswrapper[4739]: I1201 16:26:27.260877 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0630ef7a-359c-49a5-9906-b7c9e01936f7","Type":"ContainerStarted","Data":"8314bdca823e4ab828c2568203bffa822ae6fdcd6e9a30db5da7f3b4ca714af1"} Dec 01 16:26:27 crc kubenswrapper[4739]: I1201 16:26:27.262540 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"694f34a3-1680-4a0f-9320-9c0d83d957a7","Type":"ContainerStarted","Data":"cda85e24a449975ce71458cbb42fd5bfe9b3ae4ac903ecce9da34694e2d7b561"} Dec 01 16:26:27 crc kubenswrapper[4739]: I1201 16:26:27.263996 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-29c8-account-create-update-nd9g7" event={"ID":"321be3c7-476a-4d82-8bde-4a5d54f5ab9d","Type":"ContainerStarted","Data":"c5797f92b4f51634e7a03ff937825a6260c67821d162a5200ef2fe16e1e99789"} Dec 01 16:26:27 crc kubenswrapper[4739]: I1201 16:26:27.264029 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-29c8-account-create-update-nd9g7" event={"ID":"321be3c7-476a-4d82-8bde-4a5d54f5ab9d","Type":"ContainerStarted","Data":"514e5421c9fa642d3936ec8cd32884109dedf1e35fbc19ea38a099de1375ced0"} Dec 01 16:26:27 crc kubenswrapper[4739]: I1201 16:26:27.265528 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f698fbd6f-nmnlm" event={"ID":"e9e319ab-2c20-4c46-910d-980a0221a139","Type":"ContainerStarted","Data":"c93ca7e3a8547c86666650f9320fe20376fc31c4992f7b624240ba00fe24ae9c"} Dec 01 16:26:27 crc kubenswrapper[4739]: I1201 16:26:27.267685 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f84667477-69xkr" event={"ID":"a0f951fa-e09f-45ec-8110-9a7e9a1f6952","Type":"ContainerStarted","Data":"c5f81574844509023f3e28a2b21fa8368777bfc5422461fe9aad75bde2399ffd"} Dec 01 16:26:27 crc kubenswrapper[4739]: I1201 16:26:27.271951 4739 generic.go:334] "Generic (PLEG): container finished" podID="d9a7eb0a-9824-40c2-b44f-2ce4deb7604e" containerID="a1850cb88b9bec4f31329fada90766c4c363892ffb306b9eae14c4f2e986eb4a" exitCode=0 Dec 01 16:26:27 crc kubenswrapper[4739]: I1201 16:26:27.271990 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-28srx" event={"ID":"d9a7eb0a-9824-40c2-b44f-2ce4deb7604e","Type":"ContainerDied","Data":"a1850cb88b9bec4f31329fada90766c4c363892ffb306b9eae14c4f2e986eb4a"} Dec 01 16:26:27 crc kubenswrapper[4739]: I1201 16:26:27.272014 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-28srx" event={"ID":"d9a7eb0a-9824-40c2-b44f-2ce4deb7604e","Type":"ContainerStarted","Data":"1670f778a378787a3743031cefa77a928c315a821c4490d9eee5de400f06402a"} Dec 01 16:26:27 crc kubenswrapper[4739]: I1201 16:26:27.283963 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-29c8-account-create-update-nd9g7" podStartSLOduration=2.283939936 podStartE2EDuration="2.283939936s" podCreationTimestamp="2025-12-01 16:26:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:26:27.278203958 +0000 UTC m=+3089.103950062" watchObservedRunningTime="2025-12-01 16:26:27.283939936 +0000 UTC m=+3089.109686030" Dec 01 16:26:27 crc kubenswrapper[4739]: I1201 16:26:27.318612 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:26:27 crc kubenswrapper[4739]: W1201 16:26:27.370403 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7cca3f16_61f3_4fc1_8841_d3d2b168f80a.slice/crio-4bb7cf6b00b7dc3eb1ff70c4fd30053b610478b0ccaaf2623caf4514c9475412 WatchSource:0}: Error finding container 4bb7cf6b00b7dc3eb1ff70c4fd30053b610478b0ccaaf2623caf4514c9475412: Status 404 returned error can't find the container with id 4bb7cf6b00b7dc3eb1ff70c4fd30053b610478b0ccaaf2623caf4514c9475412 Dec 01 16:26:27 crc kubenswrapper[4739]: I1201 16:26:27.417949 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:26:27 crc kubenswrapper[4739]: W1201 16:26:27.428041 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bd36ab4_d8ce_4902_99d6_eb5865a70030.slice/crio-19134a13fe6753f3658e55d24514f74049b272f420c5b854f0f28653d33ea67e WatchSource:0}: Error finding container 19134a13fe6753f3658e55d24514f74049b272f420c5b854f0f28653d33ea67e: Status 404 returned error can't find the container with id 19134a13fe6753f3658e55d24514f74049b272f420c5b854f0f28653d33ea67e Dec 01 16:26:27 crc kubenswrapper[4739]: I1201 16:26:27.477276 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:26:27 crc kubenswrapper[4739]: E1201 16:26:27.477590 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.310287 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7bd36ab4-d8ce-4902-99d6-eb5865a70030","Type":"ContainerStarted","Data":"f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75"} Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.310891 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7bd36ab4-d8ce-4902-99d6-eb5865a70030","Type":"ContainerStarted","Data":"19134a13fe6753f3658e55d24514f74049b272f420c5b854f0f28653d33ea67e"} Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.311854 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f84667477-69xkr"] Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.328178 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0630ef7a-359c-49a5-9906-b7c9e01936f7","Type":"ContainerStarted","Data":"a6c5f28c3ddbd6113797104e91a48666bcc12d90e016a6a5ae0c50266822909b"} Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.328232 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0630ef7a-359c-49a5-9906-b7c9e01936f7","Type":"ContainerStarted","Data":"7873ce0ebf2c6a4dd4f4b1d54cb58dcee23773b30eae4fefa54be9ce1ae5e2b1"} Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.337695 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cca3f16-61f3-4fc1-8841-d3d2b168f80a","Type":"ContainerStarted","Data":"a2d738d5b9a39c023a6e9d7f1b2569ff4f37b83173354995b646bf67f321b6c4"} Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.337742 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cca3f16-61f3-4fc1-8841-d3d2b168f80a","Type":"ContainerStarted","Data":"4bb7cf6b00b7dc3eb1ff70c4fd30053b610478b0ccaaf2623caf4514c9475412"} Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.349622 4739 generic.go:334] "Generic (PLEG): container finished" podID="321be3c7-476a-4d82-8bde-4a5d54f5ab9d" containerID="c5797f92b4f51634e7a03ff937825a6260c67821d162a5200ef2fe16e1e99789" exitCode=0 Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.349834 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-29c8-account-create-update-nd9g7" event={"ID":"321be3c7-476a-4d82-8bde-4a5d54f5ab9d","Type":"ContainerDied","Data":"c5797f92b4f51634e7a03ff937825a6260c67821d162a5200ef2fe16e1e99789"} Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.354085 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-8675489cb4-9mz7f"] Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.356021 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.359643 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.378095 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8675489cb4-9mz7f"] Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.381344 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.25006378 podStartE2EDuration="3.379806281s" podCreationTimestamp="2025-12-01 16:26:25 +0000 UTC" firstStartedPulling="2025-12-01 16:26:26.334964281 +0000 UTC m=+3088.160710375" lastFinishedPulling="2025-12-01 16:26:27.464706782 +0000 UTC m=+3089.290452876" observedRunningTime="2025-12-01 16:26:28.345208631 +0000 UTC m=+3090.170954715" watchObservedRunningTime="2025-12-01 16:26:28.379806281 +0000 UTC m=+3090.205552395" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.495362 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-config-data\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.495478 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-horizon-tls-certs\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.495531 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-horizon-secret-key\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.495579 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-combined-ca-bundle\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.495622 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-scripts\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.495653 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-logs\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.495694 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmg78\" (UniqueName: \"kubernetes.io/projected/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-kube-api-access-kmg78\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.544680 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.556192 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f698fbd6f-nmnlm"] Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.595315 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-698d484b74-5hhp4"] Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.605855 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.610551 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-config-data\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.610615 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-horizon-tls-certs\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.610667 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-horizon-secret-key\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.610730 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-combined-ca-bundle\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.610789 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-scripts\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.610826 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-logs\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.610868 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmg78\" (UniqueName: \"kubernetes.io/projected/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-kube-api-access-kmg78\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.612906 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-config-data\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.619760 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-combined-ca-bundle\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.625217 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-horizon-secret-key\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.625796 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-scripts\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.626037 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-logs\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.647366 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.651934 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-horizon-tls-certs\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.667424 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-698d484b74-5hhp4"] Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.689067 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmg78\" (UniqueName: \"kubernetes.io/projected/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-kube-api-access-kmg78\") pod \"horizon-8675489cb4-9mz7f\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.713972 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.715222 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-horizon-tls-certs\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.715279 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-scripts\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.715376 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-logs\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.716501 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-combined-ca-bundle\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.716538 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm7jq\" (UniqueName: \"kubernetes.io/projected/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-kube-api-access-wm7jq\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.716559 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-horizon-secret-key\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.716593 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-config-data\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.818536 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-logs\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.818614 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-combined-ca-bundle\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.818651 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm7jq\" (UniqueName: \"kubernetes.io/projected/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-kube-api-access-wm7jq\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.818669 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-horizon-secret-key\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.818703 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-config-data\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.818727 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-horizon-tls-certs\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.818746 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-scripts\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.822703 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-logs\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.822881 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-scripts\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.841268 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-combined-ca-bundle\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.842627 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-config-data\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.844319 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm7jq\" (UniqueName: \"kubernetes.io/projected/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-kube-api-access-wm7jq\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.864299 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-horizon-tls-certs\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:28 crc kubenswrapper[4739]: I1201 16:26:28.866730 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/77ed99e7-001f-4dd6-9e00-5b83b7ab180f-horizon-secret-key\") pod \"horizon-698d484b74-5hhp4\" (UID: \"77ed99e7-001f-4dd6-9e00-5b83b7ab180f\") " pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.009821 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-28srx" Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.123548 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.130136 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9a7eb0a-9824-40c2-b44f-2ce4deb7604e-operator-scripts\") pod \"d9a7eb0a-9824-40c2-b44f-2ce4deb7604e\" (UID: \"d9a7eb0a-9824-40c2-b44f-2ce4deb7604e\") " Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.130199 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q42ql\" (UniqueName: \"kubernetes.io/projected/d9a7eb0a-9824-40c2-b44f-2ce4deb7604e-kube-api-access-q42ql\") pod \"d9a7eb0a-9824-40c2-b44f-2ce4deb7604e\" (UID: \"d9a7eb0a-9824-40c2-b44f-2ce4deb7604e\") " Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.131038 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9a7eb0a-9824-40c2-b44f-2ce4deb7604e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d9a7eb0a-9824-40c2-b44f-2ce4deb7604e" (UID: "d9a7eb0a-9824-40c2-b44f-2ce4deb7604e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.134013 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9a7eb0a-9824-40c2-b44f-2ce4deb7604e-kube-api-access-q42ql" (OuterVolumeSpecName: "kube-api-access-q42ql") pod "d9a7eb0a-9824-40c2-b44f-2ce4deb7604e" (UID: "d9a7eb0a-9824-40c2-b44f-2ce4deb7604e"). InnerVolumeSpecName "kube-api-access-q42ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.233299 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9a7eb0a-9824-40c2-b44f-2ce4deb7604e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.233325 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q42ql\" (UniqueName: \"kubernetes.io/projected/d9a7eb0a-9824-40c2-b44f-2ce4deb7604e-kube-api-access-q42ql\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.388106 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-28srx" event={"ID":"d9a7eb0a-9824-40c2-b44f-2ce4deb7604e","Type":"ContainerDied","Data":"1670f778a378787a3743031cefa77a928c315a821c4490d9eee5de400f06402a"} Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.388513 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1670f778a378787a3743031cefa77a928c315a821c4490d9eee5de400f06402a" Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.388122 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-28srx" Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.391551 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"694f34a3-1680-4a0f-9320-9c0d83d957a7","Type":"ContainerStarted","Data":"97a1049ef20f2b8ab953e62cb84a0af315a27fd6f538022a79d6b3a5f1deb370"} Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.472477 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8675489cb4-9mz7f"] Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.643087 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-698d484b74-5hhp4"] Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.743530 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-29c8-account-create-update-nd9g7" Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.844726 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jtzb\" (UniqueName: \"kubernetes.io/projected/321be3c7-476a-4d82-8bde-4a5d54f5ab9d-kube-api-access-2jtzb\") pod \"321be3c7-476a-4d82-8bde-4a5d54f5ab9d\" (UID: \"321be3c7-476a-4d82-8bde-4a5d54f5ab9d\") " Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.845149 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/321be3c7-476a-4d82-8bde-4a5d54f5ab9d-operator-scripts\") pod \"321be3c7-476a-4d82-8bde-4a5d54f5ab9d\" (UID: \"321be3c7-476a-4d82-8bde-4a5d54f5ab9d\") " Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.845986 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/321be3c7-476a-4d82-8bde-4a5d54f5ab9d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "321be3c7-476a-4d82-8bde-4a5d54f5ab9d" (UID: "321be3c7-476a-4d82-8bde-4a5d54f5ab9d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.852891 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/321be3c7-476a-4d82-8bde-4a5d54f5ab9d-kube-api-access-2jtzb" (OuterVolumeSpecName: "kube-api-access-2jtzb") pod "321be3c7-476a-4d82-8bde-4a5d54f5ab9d" (UID: "321be3c7-476a-4d82-8bde-4a5d54f5ab9d"). InnerVolumeSpecName "kube-api-access-2jtzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.947910 4739 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/321be3c7-476a-4d82-8bde-4a5d54f5ab9d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:29 crc kubenswrapper[4739]: I1201 16:26:29.947976 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jtzb\" (UniqueName: \"kubernetes.io/projected/321be3c7-476a-4d82-8bde-4a5d54f5ab9d-kube-api-access-2jtzb\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.412191 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"694f34a3-1680-4a0f-9320-9c0d83d957a7","Type":"ContainerStarted","Data":"314eee8707d6b47b43ea6e96f8344ffe47ceca784b4a03116eb5c9f0dd08ea36"} Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.416342 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cca3f16-61f3-4fc1-8841-d3d2b168f80a","Type":"ContainerStarted","Data":"71c75a4358d9fd9c6ce0e776cdd41a72c674ca2a03bf2d47adfe16a15f390dc7"} Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.416466 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7cca3f16-61f3-4fc1-8841-d3d2b168f80a" containerName="glance-log" containerID="cri-o://a2d738d5b9a39c023a6e9d7f1b2569ff4f37b83173354995b646bf67f321b6c4" gracePeriod=30 Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.416700 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7cca3f16-61f3-4fc1-8841-d3d2b168f80a" containerName="glance-httpd" containerID="cri-o://71c75a4358d9fd9c6ce0e776cdd41a72c674ca2a03bf2d47adfe16a15f390dc7" gracePeriod=30 Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.418104 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-29c8-account-create-update-nd9g7" event={"ID":"321be3c7-476a-4d82-8bde-4a5d54f5ab9d","Type":"ContainerDied","Data":"514e5421c9fa642d3936ec8cd32884109dedf1e35fbc19ea38a099de1375ced0"} Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.418158 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="514e5421c9fa642d3936ec8cd32884109dedf1e35fbc19ea38a099de1375ced0" Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.418227 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-29c8-account-create-update-nd9g7" Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.431866 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8675489cb4-9mz7f" event={"ID":"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7","Type":"ContainerStarted","Data":"250d778ece47f00a405232d3d1cd1bd59cd0ad8ebd1397ae09658d4a7bdd1a2e"} Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.437992 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7bd36ab4-d8ce-4902-99d6-eb5865a70030","Type":"ContainerStarted","Data":"987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813"} Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.438078 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7bd36ab4-d8ce-4902-99d6-eb5865a70030" containerName="glance-log" containerID="cri-o://f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75" gracePeriod=30 Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.438173 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7bd36ab4-d8ce-4902-99d6-eb5865a70030" containerName="glance-httpd" containerID="cri-o://987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813" gracePeriod=30 Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.450638 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-698d484b74-5hhp4" event={"ID":"77ed99e7-001f-4dd6-9e00-5b83b7ab180f","Type":"ContainerStarted","Data":"373f892865f6bd6f8808474d4b08b8392c49b45b33c59463f142d3f87793915d"} Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.463410 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=4.277227922 podStartE2EDuration="5.463372016s" podCreationTimestamp="2025-12-01 16:26:25 +0000 UTC" firstStartedPulling="2025-12-01 16:26:27.22874413 +0000 UTC m=+3089.054490224" lastFinishedPulling="2025-12-01 16:26:28.414888224 +0000 UTC m=+3090.240634318" observedRunningTime="2025-12-01 16:26:30.445001039 +0000 UTC m=+3092.270747133" watchObservedRunningTime="2025-12-01 16:26:30.463372016 +0000 UTC m=+3092.289118110" Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.506559 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.50652599 podStartE2EDuration="4.50652599s" podCreationTimestamp="2025-12-01 16:26:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:26:30.468225357 +0000 UTC m=+3092.293971451" watchObservedRunningTime="2025-12-01 16:26:30.50652599 +0000 UTC m=+3092.332272084" Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.523361 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.523322419 podStartE2EDuration="4.523322419s" podCreationTimestamp="2025-12-01 16:26:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:26:30.499952086 +0000 UTC m=+3092.325698180" watchObservedRunningTime="2025-12-01 16:26:30.523322419 +0000 UTC m=+3092.349068513" Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.629654 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:30 crc kubenswrapper[4739]: I1201 16:26:30.654725 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.118349 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.193452 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.193546 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-config-data\") pod \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.193609 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4c4tq\" (UniqueName: \"kubernetes.io/projected/7bd36ab4-d8ce-4902-99d6-eb5865a70030-kube-api-access-4c4tq\") pod \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.193685 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-combined-ca-bundle\") pod \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.193720 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-internal-tls-certs\") pod \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.193761 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bd36ab4-d8ce-4902-99d6-eb5865a70030-httpd-run\") pod \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.193784 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bd36ab4-d8ce-4902-99d6-eb5865a70030-logs\") pod \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.193817 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7bd36ab4-d8ce-4902-99d6-eb5865a70030-ceph\") pod \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.193925 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-scripts\") pod \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\" (UID: \"7bd36ab4-d8ce-4902-99d6-eb5865a70030\") " Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.195190 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bd36ab4-d8ce-4902-99d6-eb5865a70030-logs" (OuterVolumeSpecName: "logs") pod "7bd36ab4-d8ce-4902-99d6-eb5865a70030" (UID: "7bd36ab4-d8ce-4902-99d6-eb5865a70030"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.195374 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bd36ab4-d8ce-4902-99d6-eb5865a70030-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7bd36ab4-d8ce-4902-99d6-eb5865a70030" (UID: "7bd36ab4-d8ce-4902-99d6-eb5865a70030"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.204605 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bd36ab4-d8ce-4902-99d6-eb5865a70030-kube-api-access-4c4tq" (OuterVolumeSpecName: "kube-api-access-4c4tq") pod "7bd36ab4-d8ce-4902-99d6-eb5865a70030" (UID: "7bd36ab4-d8ce-4902-99d6-eb5865a70030"). InnerVolumeSpecName "kube-api-access-4c4tq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.210307 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bd36ab4-d8ce-4902-99d6-eb5865a70030-ceph" (OuterVolumeSpecName: "ceph") pod "7bd36ab4-d8ce-4902-99d6-eb5865a70030" (UID: "7bd36ab4-d8ce-4902-99d6-eb5865a70030"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.219075 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "7bd36ab4-d8ce-4902-99d6-eb5865a70030" (UID: "7bd36ab4-d8ce-4902-99d6-eb5865a70030"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.229065 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-scripts" (OuterVolumeSpecName: "scripts") pod "7bd36ab4-d8ce-4902-99d6-eb5865a70030" (UID: "7bd36ab4-d8ce-4902-99d6-eb5865a70030"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.247569 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7bd36ab4-d8ce-4902-99d6-eb5865a70030" (UID: "7bd36ab4-d8ce-4902-99d6-eb5865a70030"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.261530 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7bd36ab4-d8ce-4902-99d6-eb5865a70030" (UID: "7bd36ab4-d8ce-4902-99d6-eb5865a70030"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.279635 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-config-data" (OuterVolumeSpecName: "config-data") pod "7bd36ab4-d8ce-4902-99d6-eb5865a70030" (UID: "7bd36ab4-d8ce-4902-99d6-eb5865a70030"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.297192 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.297672 4739 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.297716 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.297762 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4c4tq\" (UniqueName: \"kubernetes.io/projected/7bd36ab4-d8ce-4902-99d6-eb5865a70030-kube-api-access-4c4tq\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.297775 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.297783 4739 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bd36ab4-d8ce-4902-99d6-eb5865a70030-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.297791 4739 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bd36ab4-d8ce-4902-99d6-eb5865a70030-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.298181 4739 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bd36ab4-d8ce-4902-99d6-eb5865a70030-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.298200 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7bd36ab4-d8ce-4902-99d6-eb5865a70030-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.324215 4739 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.324301 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-qkq9f"] Dec 01 16:26:31 crc kubenswrapper[4739]: E1201 16:26:31.324859 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9a7eb0a-9824-40c2-b44f-2ce4deb7604e" containerName="mariadb-database-create" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.324880 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9a7eb0a-9824-40c2-b44f-2ce4deb7604e" containerName="mariadb-database-create" Dec 01 16:26:31 crc kubenswrapper[4739]: E1201 16:26:31.324900 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bd36ab4-d8ce-4902-99d6-eb5865a70030" containerName="glance-log" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.324909 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bd36ab4-d8ce-4902-99d6-eb5865a70030" containerName="glance-log" Dec 01 16:26:31 crc kubenswrapper[4739]: E1201 16:26:31.324939 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="321be3c7-476a-4d82-8bde-4a5d54f5ab9d" containerName="mariadb-account-create-update" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.324947 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="321be3c7-476a-4d82-8bde-4a5d54f5ab9d" containerName="mariadb-account-create-update" Dec 01 16:26:31 crc kubenswrapper[4739]: E1201 16:26:31.324959 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bd36ab4-d8ce-4902-99d6-eb5865a70030" containerName="glance-httpd" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.324967 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bd36ab4-d8ce-4902-99d6-eb5865a70030" containerName="glance-httpd" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.325299 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bd36ab4-d8ce-4902-99d6-eb5865a70030" containerName="glance-httpd" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.325321 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9a7eb0a-9824-40c2-b44f-2ce4deb7604e" containerName="mariadb-database-create" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.325349 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="321be3c7-476a-4d82-8bde-4a5d54f5ab9d" containerName="mariadb-account-create-update" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.325362 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bd36ab4-d8ce-4902-99d6-eb5865a70030" containerName="glance-log" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.326270 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.328660 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-7x5ck" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.331278 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.335790 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-qkq9f"] Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.399361 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4t2w\" (UniqueName: \"kubernetes.io/projected/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-kube-api-access-d4t2w\") pod \"manila-db-sync-qkq9f\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.399424 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-config-data\") pod \"manila-db-sync-qkq9f\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.399444 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-job-config-data\") pod \"manila-db-sync-qkq9f\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.399522 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-combined-ca-bundle\") pod \"manila-db-sync-qkq9f\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.399646 4739 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.466258 4739 generic.go:334] "Generic (PLEG): container finished" podID="7bd36ab4-d8ce-4902-99d6-eb5865a70030" containerID="987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813" exitCode=0 Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.466288 4739 generic.go:334] "Generic (PLEG): container finished" podID="7bd36ab4-d8ce-4902-99d6-eb5865a70030" containerID="f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75" exitCode=143 Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.466323 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7bd36ab4-d8ce-4902-99d6-eb5865a70030","Type":"ContainerDied","Data":"987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813"} Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.466348 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7bd36ab4-d8ce-4902-99d6-eb5865a70030","Type":"ContainerDied","Data":"f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75"} Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.466358 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7bd36ab4-d8ce-4902-99d6-eb5865a70030","Type":"ContainerDied","Data":"19134a13fe6753f3658e55d24514f74049b272f420c5b854f0f28653d33ea67e"} Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.466372 4739 scope.go:117] "RemoveContainer" containerID="987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.466380 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.471966 4739 generic.go:334] "Generic (PLEG): container finished" podID="7cca3f16-61f3-4fc1-8841-d3d2b168f80a" containerID="71c75a4358d9fd9c6ce0e776cdd41a72c674ca2a03bf2d47adfe16a15f390dc7" exitCode=0 Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.472026 4739 generic.go:334] "Generic (PLEG): container finished" podID="7cca3f16-61f3-4fc1-8841-d3d2b168f80a" containerID="a2d738d5b9a39c023a6e9d7f1b2569ff4f37b83173354995b646bf67f321b6c4" exitCode=143 Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.472029 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cca3f16-61f3-4fc1-8841-d3d2b168f80a","Type":"ContainerDied","Data":"71c75a4358d9fd9c6ce0e776cdd41a72c674ca2a03bf2d47adfe16a15f390dc7"} Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.472069 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cca3f16-61f3-4fc1-8841-d3d2b168f80a","Type":"ContainerDied","Data":"a2d738d5b9a39c023a6e9d7f1b2569ff4f37b83173354995b646bf67f321b6c4"} Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.502571 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4t2w\" (UniqueName: \"kubernetes.io/projected/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-kube-api-access-d4t2w\") pod \"manila-db-sync-qkq9f\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.502613 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-config-data\") pod \"manila-db-sync-qkq9f\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.502639 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-job-config-data\") pod \"manila-db-sync-qkq9f\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.502767 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-combined-ca-bundle\") pod \"manila-db-sync-qkq9f\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.512546 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-config-data\") pod \"manila-db-sync-qkq9f\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.520287 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-combined-ca-bundle\") pod \"manila-db-sync-qkq9f\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.526112 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-job-config-data\") pod \"manila-db-sync-qkq9f\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.531799 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.538844 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4t2w\" (UniqueName: \"kubernetes.io/projected/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-kube-api-access-d4t2w\") pod \"manila-db-sync-qkq9f\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.551315 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.555075 4739 scope.go:117] "RemoveContainer" containerID="f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.574828 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.589017 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.591550 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.591801 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.629906 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.647428 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.673751 4739 scope.go:117] "RemoveContainer" containerID="987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813" Dec 01 16:26:31 crc kubenswrapper[4739]: E1201 16:26:31.681192 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813\": container with ID starting with 987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813 not found: ID does not exist" containerID="987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.681275 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813"} err="failed to get container status \"987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813\": rpc error: code = NotFound desc = could not find container \"987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813\": container with ID starting with 987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813 not found: ID does not exist" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.681298 4739 scope.go:117] "RemoveContainer" containerID="f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75" Dec 01 16:26:31 crc kubenswrapper[4739]: E1201 16:26:31.682081 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75\": container with ID starting with f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75 not found: ID does not exist" containerID="f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.682109 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75"} err="failed to get container status \"f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75\": rpc error: code = NotFound desc = could not find container \"f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75\": container with ID starting with f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75 not found: ID does not exist" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.682121 4739 scope.go:117] "RemoveContainer" containerID="987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.683351 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813"} err="failed to get container status \"987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813\": rpc error: code = NotFound desc = could not find container \"987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813\": container with ID starting with 987bd26a96e89f7ff64d3af6368a03ecebf8bf58d2274a3fdd74ad97a8d64813 not found: ID does not exist" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.683376 4739 scope.go:117] "RemoveContainer" containerID="f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.684687 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75"} err="failed to get container status \"f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75\": rpc error: code = NotFound desc = could not find container \"f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75\": container with ID starting with f41b451ece93c9fdd4b05ea7da670cbf49ef4f5ec80954dfcc0a6dc133000b75 not found: ID does not exist" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.706831 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.706878 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/201539c2-475c-4512-912b-729e12b1ba88-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.706912 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/201539c2-475c-4512-912b-729e12b1ba88-scripts\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.707110 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/201539c2-475c-4512-912b-729e12b1ba88-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.707133 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/201539c2-475c-4512-912b-729e12b1ba88-config-data\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.707147 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/201539c2-475c-4512-912b-729e12b1ba88-logs\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.707550 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/201539c2-475c-4512-912b-729e12b1ba88-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.707606 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/201539c2-475c-4512-912b-729e12b1ba88-ceph\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.707654 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84lkh\" (UniqueName: \"kubernetes.io/projected/201539c2-475c-4512-912b-729e12b1ba88-kube-api-access-84lkh\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.809119 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.809484 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/201539c2-475c-4512-912b-729e12b1ba88-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.809525 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/201539c2-475c-4512-912b-729e12b1ba88-scripts\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.809557 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/201539c2-475c-4512-912b-729e12b1ba88-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.809580 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/201539c2-475c-4512-912b-729e12b1ba88-config-data\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.809597 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/201539c2-475c-4512-912b-729e12b1ba88-logs\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.809639 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/201539c2-475c-4512-912b-729e12b1ba88-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.809688 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/201539c2-475c-4512-912b-729e12b1ba88-ceph\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.809733 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84lkh\" (UniqueName: \"kubernetes.io/projected/201539c2-475c-4512-912b-729e12b1ba88-kube-api-access-84lkh\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.811162 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/201539c2-475c-4512-912b-729e12b1ba88-logs\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.811510 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/201539c2-475c-4512-912b-729e12b1ba88-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.812553 4739 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.820172 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/201539c2-475c-4512-912b-729e12b1ba88-scripts\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.821207 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/201539c2-475c-4512-912b-729e12b1ba88-config-data\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.825181 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/201539c2-475c-4512-912b-729e12b1ba88-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.828162 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/201539c2-475c-4512-912b-729e12b1ba88-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.834031 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84lkh\" (UniqueName: \"kubernetes.io/projected/201539c2-475c-4512-912b-729e12b1ba88-kube-api-access-84lkh\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.836606 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/201539c2-475c-4512-912b-729e12b1ba88-ceph\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.846505 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"201539c2-475c-4512-912b-729e12b1ba88\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: E1201 16:26:31.892731 4739 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bd36ab4_d8ce_4902_99d6_eb5865a70030.slice\": RecentStats: unable to find data in memory cache]" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.940800 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:26:31 crc kubenswrapper[4739]: I1201 16:26:31.967356 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.013369 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-public-tls-certs\") pod \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.013894 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-combined-ca-bundle\") pod \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.013970 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-ceph\") pod \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.013993 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.014026 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bw2g\" (UniqueName: \"kubernetes.io/projected/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-kube-api-access-6bw2g\") pod \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.014068 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-config-data\") pod \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.014142 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-scripts\") pod \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.014173 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-httpd-run\") pod \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.014232 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-logs\") pod \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\" (UID: \"7cca3f16-61f3-4fc1-8841-d3d2b168f80a\") " Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.014886 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-logs" (OuterVolumeSpecName: "logs") pod "7cca3f16-61f3-4fc1-8841-d3d2b168f80a" (UID: "7cca3f16-61f3-4fc1-8841-d3d2b168f80a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.018958 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7cca3f16-61f3-4fc1-8841-d3d2b168f80a" (UID: "7cca3f16-61f3-4fc1-8841-d3d2b168f80a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.029179 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-kube-api-access-6bw2g" (OuterVolumeSpecName: "kube-api-access-6bw2g") pod "7cca3f16-61f3-4fc1-8841-d3d2b168f80a" (UID: "7cca3f16-61f3-4fc1-8841-d3d2b168f80a"). InnerVolumeSpecName "kube-api-access-6bw2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.029694 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-scripts" (OuterVolumeSpecName: "scripts") pod "7cca3f16-61f3-4fc1-8841-d3d2b168f80a" (UID: "7cca3f16-61f3-4fc1-8841-d3d2b168f80a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.029935 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-ceph" (OuterVolumeSpecName: "ceph") pod "7cca3f16-61f3-4fc1-8841-d3d2b168f80a" (UID: "7cca3f16-61f3-4fc1-8841-d3d2b168f80a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.030350 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "7cca3f16-61f3-4fc1-8841-d3d2b168f80a" (UID: "7cca3f16-61f3-4fc1-8841-d3d2b168f80a"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.079946 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7cca3f16-61f3-4fc1-8841-d3d2b168f80a" (UID: "7cca3f16-61f3-4fc1-8841-d3d2b168f80a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.083697 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7cca3f16-61f3-4fc1-8841-d3d2b168f80a" (UID: "7cca3f16-61f3-4fc1-8841-d3d2b168f80a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.101169 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-config-data" (OuterVolumeSpecName: "config-data") pod "7cca3f16-61f3-4fc1-8841-d3d2b168f80a" (UID: "7cca3f16-61f3-4fc1-8841-d3d2b168f80a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.116727 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.116761 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.116809 4739 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.116820 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bw2g\" (UniqueName: \"kubernetes.io/projected/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-kube-api-access-6bw2g\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.116831 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.116839 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.116847 4739 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.116854 4739 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.116862 4739 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cca3f16-61f3-4fc1-8841-d3d2b168f80a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.136164 4739 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.221211 4739 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.324197 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-qkq9f"] Dec 01 16:26:32 crc kubenswrapper[4739]: W1201 16:26:32.334537 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca6268bc_3fab_4df5_a86e_f2060ab9abb8.slice/crio-9cf7246478534b1461bec71ebad315bfca970533af0681cffc60ce2cd24ba866 WatchSource:0}: Error finding container 9cf7246478534b1461bec71ebad315bfca970533af0681cffc60ce2cd24ba866: Status 404 returned error can't find the container with id 9cf7246478534b1461bec71ebad315bfca970533af0681cffc60ce2cd24ba866 Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.509270 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bd36ab4-d8ce-4902-99d6-eb5865a70030" path="/var/lib/kubelet/pods/7bd36ab4-d8ce-4902-99d6-eb5865a70030/volumes" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.510714 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-qkq9f" event={"ID":"ca6268bc-3fab-4df5-a86e-f2060ab9abb8","Type":"ContainerStarted","Data":"9cf7246478534b1461bec71ebad315bfca970533af0681cffc60ce2cd24ba866"} Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.520849 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cca3f16-61f3-4fc1-8841-d3d2b168f80a","Type":"ContainerDied","Data":"4bb7cf6b00b7dc3eb1ff70c4fd30053b610478b0ccaaf2623caf4514c9475412"} Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.520933 4739 scope.go:117] "RemoveContainer" containerID="71c75a4358d9fd9c6ce0e776cdd41a72c674ca2a03bf2d47adfe16a15f390dc7" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.521614 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.569019 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.583755 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.600577 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.610324 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:26:32 crc kubenswrapper[4739]: E1201 16:26:32.613384 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cca3f16-61f3-4fc1-8841-d3d2b168f80a" containerName="glance-httpd" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.613404 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cca3f16-61f3-4fc1-8841-d3d2b168f80a" containerName="glance-httpd" Dec 01 16:26:32 crc kubenswrapper[4739]: E1201 16:26:32.613428 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cca3f16-61f3-4fc1-8841-d3d2b168f80a" containerName="glance-log" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.616162 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cca3f16-61f3-4fc1-8841-d3d2b168f80a" containerName="glance-log" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.616788 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cca3f16-61f3-4fc1-8841-d3d2b168f80a" containerName="glance-httpd" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.616823 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cca3f16-61f3-4fc1-8841-d3d2b168f80a" containerName="glance-log" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.618671 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.623329 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.623923 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.624150 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.732361 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.732509 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-ceph\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.732543 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.732657 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v74f\" (UniqueName: \"kubernetes.io/projected/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-kube-api-access-7v74f\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.732701 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.732754 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.732805 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-config-data\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.732894 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-scripts\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.733002 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-logs\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.835128 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.835542 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-config-data\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.835597 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-scripts\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.835665 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-logs\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.835744 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.835790 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-ceph\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.835819 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.835850 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v74f\" (UniqueName: \"kubernetes.io/projected/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-kube-api-access-7v74f\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.835887 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.839020 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-logs\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.839296 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.839432 4739 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.843389 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.844063 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.857282 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-scripts\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.859815 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-ceph\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.860772 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v74f\" (UniqueName: \"kubernetes.io/projected/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-kube-api-access-7v74f\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.861072 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d0ae7f-f45b-40a0-870b-f1d7c6252e45-config-data\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.870736 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"47d0ae7f-f45b-40a0-870b-f1d7c6252e45\") " pod="openstack/glance-default-external-api-0" Dec 01 16:26:32 crc kubenswrapper[4739]: I1201 16:26:32.982974 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:26:34 crc kubenswrapper[4739]: I1201 16:26:34.490683 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cca3f16-61f3-4fc1-8841-d3d2b168f80a" path="/var/lib/kubelet/pods/7cca3f16-61f3-4fc1-8841-d3d2b168f80a/volumes" Dec 01 16:26:35 crc kubenswrapper[4739]: I1201 16:26:35.821655 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Dec 01 16:26:35 crc kubenswrapper[4739]: I1201 16:26:35.829975 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 01 16:26:37 crc kubenswrapper[4739]: I1201 16:26:37.836736 4739 scope.go:117] "RemoveContainer" containerID="a2d738d5b9a39c023a6e9d7f1b2569ff4f37b83173354995b646bf67f321b6c4" Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.473555 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:26:38 crc kubenswrapper[4739]: W1201 16:26:38.493136 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47d0ae7f_f45b_40a0_870b_f1d7c6252e45.slice/crio-58ef9b164b4349d1800d6752669eef7c7d0bc121f3266d1d6edbcdcab345c7df WatchSource:0}: Error finding container 58ef9b164b4349d1800d6752669eef7c7d0bc121f3266d1d6edbcdcab345c7df: Status 404 returned error can't find the container with id 58ef9b164b4349d1800d6752669eef7c7d0bc121f3266d1d6edbcdcab345c7df Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.601740 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-698d484b74-5hhp4" event={"ID":"77ed99e7-001f-4dd6-9e00-5b83b7ab180f","Type":"ContainerStarted","Data":"7ecf885284e94cebe6c101d9a73da3077ad3c3045f57aaf68ff84cc3e2bfd2cd"} Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.604040 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47d0ae7f-f45b-40a0-870b-f1d7c6252e45","Type":"ContainerStarted","Data":"58ef9b164b4349d1800d6752669eef7c7d0bc121f3266d1d6edbcdcab345c7df"} Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.610094 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8675489cb4-9mz7f" event={"ID":"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7","Type":"ContainerStarted","Data":"c30f8e2e727a49a2094463f0e8cf3a19335dd473e4173694417d8ca96e055330"} Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.610151 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8675489cb4-9mz7f" event={"ID":"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7","Type":"ContainerStarted","Data":"82e1bf034e64723a1b9fead779cd64592427fad19be93d92b9185d673b6e38ed"} Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.613769 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"201539c2-475c-4512-912b-729e12b1ba88","Type":"ContainerStarted","Data":"938fbe3453f4fea945b2546753e4ae401ab15eff55ddad43983d27fa3bc9cca4"} Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.615725 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f698fbd6f-nmnlm" event={"ID":"e9e319ab-2c20-4c46-910d-980a0221a139","Type":"ContainerStarted","Data":"4d5f9636692ba65ab6b56afe3e47785bfa8a57ec307a0eeb7cb1d1e1eb901780"} Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.615753 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f698fbd6f-nmnlm" event={"ID":"e9e319ab-2c20-4c46-910d-980a0221a139","Type":"ContainerStarted","Data":"f467346191e267478cd2ad6de53a7393c4b56dbedc0348a05186340f1d422fd5"} Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.615867 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f698fbd6f-nmnlm" podUID="e9e319ab-2c20-4c46-910d-980a0221a139" containerName="horizon-log" containerID="cri-o://f467346191e267478cd2ad6de53a7393c4b56dbedc0348a05186340f1d422fd5" gracePeriod=30 Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.616194 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f698fbd6f-nmnlm" podUID="e9e319ab-2c20-4c46-910d-980a0221a139" containerName="horizon" containerID="cri-o://4d5f9636692ba65ab6b56afe3e47785bfa8a57ec307a0eeb7cb1d1e1eb901780" gracePeriod=30 Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.623913 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f84667477-69xkr" event={"ID":"a0f951fa-e09f-45ec-8110-9a7e9a1f6952","Type":"ContainerStarted","Data":"101e1b3a3b5f4f8110f6505354518dcc71eaa944706c9e6a66750759257ee253"} Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.623961 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f84667477-69xkr" event={"ID":"a0f951fa-e09f-45ec-8110-9a7e9a1f6952","Type":"ContainerStarted","Data":"24d2e3b604ad8720e2fab2b7f8d6f358e4b3560908a5583d4071035547398cff"} Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.624065 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f84667477-69xkr" podUID="a0f951fa-e09f-45ec-8110-9a7e9a1f6952" containerName="horizon-log" containerID="cri-o://24d2e3b604ad8720e2fab2b7f8d6f358e4b3560908a5583d4071035547398cff" gracePeriod=30 Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.624308 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f84667477-69xkr" podUID="a0f951fa-e09f-45ec-8110-9a7e9a1f6952" containerName="horizon" containerID="cri-o://101e1b3a3b5f4f8110f6505354518dcc71eaa944706c9e6a66750759257ee253" gracePeriod=30 Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.655113 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-8675489cb4-9mz7f" podStartSLOduration=2.130247782 podStartE2EDuration="10.655085575s" podCreationTimestamp="2025-12-01 16:26:28 +0000 UTC" firstStartedPulling="2025-12-01 16:26:29.498079157 +0000 UTC m=+3091.323825251" lastFinishedPulling="2025-12-01 16:26:38.02291694 +0000 UTC m=+3099.848663044" observedRunningTime="2025-12-01 16:26:38.632092756 +0000 UTC m=+3100.457838850" watchObservedRunningTime="2025-12-01 16:26:38.655085575 +0000 UTC m=+3100.480831669" Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.670693 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7f84667477-69xkr" podStartSLOduration=2.631773624 podStartE2EDuration="13.670670027s" podCreationTimestamp="2025-12-01 16:26:25 +0000 UTC" firstStartedPulling="2025-12-01 16:26:26.981389346 +0000 UTC m=+3088.807135440" lastFinishedPulling="2025-12-01 16:26:38.020285729 +0000 UTC m=+3099.846031843" observedRunningTime="2025-12-01 16:26:38.647922574 +0000 UTC m=+3100.473668668" watchObservedRunningTime="2025-12-01 16:26:38.670670027 +0000 UTC m=+3100.496416131" Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.688327 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7f698fbd6f-nmnlm" podStartSLOduration=1.692721108 podStartE2EDuration="12.688307673s" podCreationTimestamp="2025-12-01 16:26:26 +0000 UTC" firstStartedPulling="2025-12-01 16:26:27.026552012 +0000 UTC m=+3088.852298106" lastFinishedPulling="2025-12-01 16:26:38.022138577 +0000 UTC m=+3099.847884671" observedRunningTime="2025-12-01 16:26:38.668668296 +0000 UTC m=+3100.494414390" watchObservedRunningTime="2025-12-01 16:26:38.688307673 +0000 UTC m=+3100.514053767" Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.715880 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:38 crc kubenswrapper[4739]: I1201 16:26:38.716295 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:26:39 crc kubenswrapper[4739]: I1201 16:26:39.639296 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"201539c2-475c-4512-912b-729e12b1ba88","Type":"ContainerStarted","Data":"69c13805f3d5e23001756cdecc0b1f19d691d203adac4f6b27ba489d2aeb271f"} Dec 01 16:26:39 crc kubenswrapper[4739]: I1201 16:26:39.639653 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"201539c2-475c-4512-912b-729e12b1ba88","Type":"ContainerStarted","Data":"021a0fdcab8d2cc2a17fda49694361e858793e3e5f04d92123b523e1d6266ebd"} Dec 01 16:26:39 crc kubenswrapper[4739]: I1201 16:26:39.642937 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-698d484b74-5hhp4" event={"ID":"77ed99e7-001f-4dd6-9e00-5b83b7ab180f","Type":"ContainerStarted","Data":"360e74e36cac7f1b92ad3cbb64acfcbb470e0f209d7e8ec5fb734d60b81a979e"} Dec 01 16:26:39 crc kubenswrapper[4739]: I1201 16:26:39.650159 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47d0ae7f-f45b-40a0-870b-f1d7c6252e45","Type":"ContainerStarted","Data":"71900ea20a2cc5557d9894e065f19c35ed8f9a9c39b4ad90c347f34bf2594994"} Dec 01 16:26:39 crc kubenswrapper[4739]: I1201 16:26:39.650206 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47d0ae7f-f45b-40a0-870b-f1d7c6252e45","Type":"ContainerStarted","Data":"bf26190a967f2623508739e50b9e9bacd355a4dcce85d6443984e329c8384685"} Dec 01 16:26:39 crc kubenswrapper[4739]: I1201 16:26:39.661482 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.661463205 podStartE2EDuration="8.661463205s" podCreationTimestamp="2025-12-01 16:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:26:39.660714111 +0000 UTC m=+3101.486460205" watchObservedRunningTime="2025-12-01 16:26:39.661463205 +0000 UTC m=+3101.487209829" Dec 01 16:26:39 crc kubenswrapper[4739]: I1201 16:26:39.679663 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-698d484b74-5hhp4" podStartSLOduration=3.313029002 podStartE2EDuration="11.679643976s" podCreationTimestamp="2025-12-01 16:26:28 +0000 UTC" firstStartedPulling="2025-12-01 16:26:29.655643216 +0000 UTC m=+3091.481389310" lastFinishedPulling="2025-12-01 16:26:38.02225818 +0000 UTC m=+3099.848004284" observedRunningTime="2025-12-01 16:26:39.679045108 +0000 UTC m=+3101.504791202" watchObservedRunningTime="2025-12-01 16:26:39.679643976 +0000 UTC m=+3101.505390070" Dec 01 16:26:39 crc kubenswrapper[4739]: I1201 16:26:39.706078 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.706061713 podStartE2EDuration="7.706061713s" podCreationTimestamp="2025-12-01 16:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:26:39.700289674 +0000 UTC m=+3101.526035768" watchObservedRunningTime="2025-12-01 16:26:39.706061713 +0000 UTC m=+3101.531807807" Dec 01 16:26:41 crc kubenswrapper[4739]: I1201 16:26:41.477414 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:26:41 crc kubenswrapper[4739]: E1201 16:26:41.478278 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:26:41 crc kubenswrapper[4739]: I1201 16:26:41.968178 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:41 crc kubenswrapper[4739]: I1201 16:26:41.969653 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:42 crc kubenswrapper[4739]: I1201 16:26:42.037460 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:42 crc kubenswrapper[4739]: I1201 16:26:42.042579 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:42 crc kubenswrapper[4739]: I1201 16:26:42.686802 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:42 crc kubenswrapper[4739]: I1201 16:26:42.689687 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:42 crc kubenswrapper[4739]: I1201 16:26:42.984243 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 16:26:42 crc kubenswrapper[4739]: I1201 16:26:42.984296 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 16:26:43 crc kubenswrapper[4739]: I1201 16:26:43.035587 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 16:26:43 crc kubenswrapper[4739]: I1201 16:26:43.044846 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 16:26:43 crc kubenswrapper[4739]: I1201 16:26:43.695170 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-qkq9f" event={"ID":"ca6268bc-3fab-4df5-a86e-f2060ab9abb8","Type":"ContainerStarted","Data":"db11f88319bf52fb097b2224554764287f16f05f3e94485dad5a65ede39487f4"} Dec 01 16:26:43 crc kubenswrapper[4739]: I1201 16:26:43.695750 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 16:26:43 crc kubenswrapper[4739]: I1201 16:26:43.695769 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 16:26:43 crc kubenswrapper[4739]: I1201 16:26:43.730403 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-qkq9f" podStartSLOduration=2.533741857 podStartE2EDuration="12.730377962s" podCreationTimestamp="2025-12-01 16:26:31 +0000 UTC" firstStartedPulling="2025-12-01 16:26:32.337190461 +0000 UTC m=+3094.162936555" lastFinishedPulling="2025-12-01 16:26:42.533826566 +0000 UTC m=+3104.359572660" observedRunningTime="2025-12-01 16:26:43.725739628 +0000 UTC m=+3105.551485722" watchObservedRunningTime="2025-12-01 16:26:43.730377962 +0000 UTC m=+3105.556124056" Dec 01 16:26:46 crc kubenswrapper[4739]: I1201 16:26:46.269167 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:26:46 crc kubenswrapper[4739]: I1201 16:26:46.277794 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:46 crc kubenswrapper[4739]: I1201 16:26:46.456837 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:26:46 crc kubenswrapper[4739]: I1201 16:26:46.560813 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 16:26:46 crc kubenswrapper[4739]: I1201 16:26:46.729857 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 16:26:46 crc kubenswrapper[4739]: I1201 16:26:46.815904 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 16:26:48 crc kubenswrapper[4739]: I1201 16:26:48.717494 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-8675489cb4-9mz7f" podUID="0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.240:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.240:8443: connect: connection refused" Dec 01 16:26:49 crc kubenswrapper[4739]: I1201 16:26:49.123754 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:49 crc kubenswrapper[4739]: I1201 16:26:49.123841 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:26:49 crc kubenswrapper[4739]: I1201 16:26:49.126350 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-698d484b74-5hhp4" podUID="77ed99e7-001f-4dd6-9e00-5b83b7ab180f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.241:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.241:8443: connect: connection refused" Dec 01 16:26:53 crc kubenswrapper[4739]: I1201 16:26:53.793790 4739 generic.go:334] "Generic (PLEG): container finished" podID="ca6268bc-3fab-4df5-a86e-f2060ab9abb8" containerID="db11f88319bf52fb097b2224554764287f16f05f3e94485dad5a65ede39487f4" exitCode=0 Dec 01 16:26:53 crc kubenswrapper[4739]: I1201 16:26:53.793911 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-qkq9f" event={"ID":"ca6268bc-3fab-4df5-a86e-f2060ab9abb8","Type":"ContainerDied","Data":"db11f88319bf52fb097b2224554764287f16f05f3e94485dad5a65ede39487f4"} Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.233044 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.300873 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-combined-ca-bundle\") pod \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.301107 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-config-data\") pod \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.301207 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-job-config-data\") pod \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.301258 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4t2w\" (UniqueName: \"kubernetes.io/projected/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-kube-api-access-d4t2w\") pod \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\" (UID: \"ca6268bc-3fab-4df5-a86e-f2060ab9abb8\") " Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.308691 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "ca6268bc-3fab-4df5-a86e-f2060ab9abb8" (UID: "ca6268bc-3fab-4df5-a86e-f2060ab9abb8"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.318124 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-config-data" (OuterVolumeSpecName: "config-data") pod "ca6268bc-3fab-4df5-a86e-f2060ab9abb8" (UID: "ca6268bc-3fab-4df5-a86e-f2060ab9abb8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.327223 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-kube-api-access-d4t2w" (OuterVolumeSpecName: "kube-api-access-d4t2w") pod "ca6268bc-3fab-4df5-a86e-f2060ab9abb8" (UID: "ca6268bc-3fab-4df5-a86e-f2060ab9abb8"). InnerVolumeSpecName "kube-api-access-d4t2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.349472 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca6268bc-3fab-4df5-a86e-f2060ab9abb8" (UID: "ca6268bc-3fab-4df5-a86e-f2060ab9abb8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.404927 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.404959 4739 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-job-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.404975 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4t2w\" (UniqueName: \"kubernetes.io/projected/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-kube-api-access-d4t2w\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.404988 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca6268bc-3fab-4df5-a86e-f2060ab9abb8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.478631 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:26:55 crc kubenswrapper[4739]: E1201 16:26:55.479092 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.825279 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-qkq9f" event={"ID":"ca6268bc-3fab-4df5-a86e-f2060ab9abb8","Type":"ContainerDied","Data":"9cf7246478534b1461bec71ebad315bfca970533af0681cffc60ce2cd24ba866"} Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.825373 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cf7246478534b1461bec71ebad315bfca970533af0681cffc60ce2cd24ba866" Dec 01 16:26:55 crc kubenswrapper[4739]: I1201 16:26:55.825315 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-qkq9f" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.124309 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 16:26:56 crc kubenswrapper[4739]: E1201 16:26:56.124877 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca6268bc-3fab-4df5-a86e-f2060ab9abb8" containerName="manila-db-sync" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.124899 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca6268bc-3fab-4df5-a86e-f2060ab9abb8" containerName="manila-db-sync" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.125163 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca6268bc-3fab-4df5-a86e-f2060ab9abb8" containerName="manila-db-sync" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.126513 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.130246 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.130683 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.130878 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.131177 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-7x5ck" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.146630 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.217129 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.218701 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.222674 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.228744 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3930d334-34db-4729-9064-140ba9d413f4-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.228811 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.228835 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.228852 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-scripts\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.228885 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dhqd\" (UniqueName: \"kubernetes.io/projected/5b0dcc39-d1cc-454a-a403-70d47279541c-kube-api-access-7dhqd\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.228957 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8htn\" (UniqueName: \"kubernetes.io/projected/3930d334-34db-4729-9064-140ba9d413f4-kube-api-access-q8htn\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.228978 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-scripts\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.229037 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.229070 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-config-data\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.229103 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3930d334-34db-4729-9064-140ba9d413f4-ceph\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.229136 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-config-data\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.229215 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.229261 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b0dcc39-d1cc-454a-a403-70d47279541c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.229289 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3930d334-34db-4729-9064-140ba9d413f4-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.264651 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.331310 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b0dcc39-d1cc-454a-a403-70d47279541c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.331647 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3930d334-34db-4729-9064-140ba9d413f4-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.331694 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3930d334-34db-4729-9064-140ba9d413f4-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.331732 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.331753 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.331768 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-scripts\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.331788 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dhqd\" (UniqueName: \"kubernetes.io/projected/5b0dcc39-d1cc-454a-a403-70d47279541c-kube-api-access-7dhqd\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.331831 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8htn\" (UniqueName: \"kubernetes.io/projected/3930d334-34db-4729-9064-140ba9d413f4-kube-api-access-q8htn\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.331849 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-scripts\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.331907 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.331935 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-config-data\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.331967 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3930d334-34db-4729-9064-140ba9d413f4-ceph\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.331994 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-config-data\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.332038 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.333020 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3930d334-34db-4729-9064-140ba9d413f4-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.333645 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b0dcc39-d1cc-454a-a403-70d47279541c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.333645 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3930d334-34db-4729-9064-140ba9d413f4-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.340831 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76b5fdb995-pvgg8"] Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.341328 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-config-data\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.342354 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.343794 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-scripts\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.344696 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.352762 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76b5fdb995-pvgg8"] Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.354029 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-scripts\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.354534 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-config-data\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.354905 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3930d334-34db-4729-9064-140ba9d413f4-ceph\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.362063 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.369615 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dhqd\" (UniqueName: \"kubernetes.io/projected/5b0dcc39-d1cc-454a-a403-70d47279541c-kube-api-access-7dhqd\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.369944 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8htn\" (UniqueName: \"kubernetes.io/projected/3930d334-34db-4729-9064-140ba9d413f4-kube-api-access-q8htn\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.371994 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.372140 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.434953 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f27b159a-6738-43d0-8ba1-041cc46170dc-openstack-edpm-ipam\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.435018 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwwtw\" (UniqueName: \"kubernetes.io/projected/f27b159a-6738-43d0-8ba1-041cc46170dc-kube-api-access-dwwtw\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.435084 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f27b159a-6738-43d0-8ba1-041cc46170dc-dns-svc\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.435126 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f27b159a-6738-43d0-8ba1-041cc46170dc-config\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.435186 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f27b159a-6738-43d0-8ba1-041cc46170dc-ovsdbserver-nb\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.435235 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f27b159a-6738-43d0-8ba1-041cc46170dc-ovsdbserver-sb\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.445433 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.538012 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f27b159a-6738-43d0-8ba1-041cc46170dc-config\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.538169 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f27b159a-6738-43d0-8ba1-041cc46170dc-ovsdbserver-nb\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.538220 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f27b159a-6738-43d0-8ba1-041cc46170dc-ovsdbserver-sb\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.538246 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f27b159a-6738-43d0-8ba1-041cc46170dc-openstack-edpm-ipam\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.538281 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwwtw\" (UniqueName: \"kubernetes.io/projected/f27b159a-6738-43d0-8ba1-041cc46170dc-kube-api-access-dwwtw\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.538346 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f27b159a-6738-43d0-8ba1-041cc46170dc-dns-svc\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.539203 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f27b159a-6738-43d0-8ba1-041cc46170dc-config\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.540496 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f27b159a-6738-43d0-8ba1-041cc46170dc-dns-svc\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.541403 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f27b159a-6738-43d0-8ba1-041cc46170dc-openstack-edpm-ipam\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.541403 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f27b159a-6738-43d0-8ba1-041cc46170dc-ovsdbserver-sb\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.541807 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f27b159a-6738-43d0-8ba1-041cc46170dc-ovsdbserver-nb\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.556939 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.574547 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.576508 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.580788 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.585484 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwwtw\" (UniqueName: \"kubernetes.io/projected/f27b159a-6738-43d0-8ba1-041cc46170dc-kube-api-access-dwwtw\") pod \"dnsmasq-dns-76b5fdb995-pvgg8\" (UID: \"f27b159a-6738-43d0-8ba1-041cc46170dc\") " pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.650544 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.744536 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-config-data\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.744804 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/57dfdfae-381a-46ca-921c-2528f1c11d05-etc-machine-id\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.744894 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-scripts\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.744939 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.745003 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk29c\" (UniqueName: \"kubernetes.io/projected/57dfdfae-381a-46ca-921c-2528f1c11d05-kube-api-access-gk29c\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.745063 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57dfdfae-381a-46ca-921c-2528f1c11d05-logs\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.745113 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-config-data-custom\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.849973 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.850041 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk29c\" (UniqueName: \"kubernetes.io/projected/57dfdfae-381a-46ca-921c-2528f1c11d05-kube-api-access-gk29c\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.850059 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57dfdfae-381a-46ca-921c-2528f1c11d05-logs\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.850079 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-config-data-custom\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.850166 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-config-data\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.850183 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/57dfdfae-381a-46ca-921c-2528f1c11d05-etc-machine-id\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.850226 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-scripts\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.850960 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57dfdfae-381a-46ca-921c-2528f1c11d05-logs\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.849971 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.851741 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/57dfdfae-381a-46ca-921c-2528f1c11d05-etc-machine-id\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.866817 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-scripts\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.869024 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-config-data\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.870958 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-config-data-custom\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.878273 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk29c\" (UniqueName: \"kubernetes.io/projected/57dfdfae-381a-46ca-921c-2528f1c11d05-kube-api-access-gk29c\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.884939 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " pod="openstack/manila-api-0" Dec 01 16:26:56 crc kubenswrapper[4739]: I1201 16:26:56.981932 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 01 16:26:57 crc kubenswrapper[4739]: I1201 16:26:57.036688 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 16:26:57 crc kubenswrapper[4739]: W1201 16:26:57.047760 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b0dcc39_d1cc_454a_a403_70d47279541c.slice/crio-a571b6c958638eafcef691e477afcd94f108497774c3aff88d1c6a7bba8d7d4b WatchSource:0}: Error finding container a571b6c958638eafcef691e477afcd94f108497774c3aff88d1c6a7bba8d7d4b: Status 404 returned error can't find the container with id a571b6c958638eafcef691e477afcd94f108497774c3aff88d1c6a7bba8d7d4b Dec 01 16:26:57 crc kubenswrapper[4739]: I1201 16:26:57.250532 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 16:26:57 crc kubenswrapper[4739]: W1201 16:26:57.352939 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf27b159a_6738_43d0_8ba1_041cc46170dc.slice/crio-88c349ee704701af94c4bcc37e38a7cc28bf381ad476f935365663c48a5eea0a WatchSource:0}: Error finding container 88c349ee704701af94c4bcc37e38a7cc28bf381ad476f935365663c48a5eea0a: Status 404 returned error can't find the container with id 88c349ee704701af94c4bcc37e38a7cc28bf381ad476f935365663c48a5eea0a Dec 01 16:26:57 crc kubenswrapper[4739]: I1201 16:26:57.353991 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76b5fdb995-pvgg8"] Dec 01 16:26:57 crc kubenswrapper[4739]: I1201 16:26:57.624337 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 01 16:26:57 crc kubenswrapper[4739]: I1201 16:26:57.863100 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"57dfdfae-381a-46ca-921c-2528f1c11d05","Type":"ContainerStarted","Data":"626b7ebf51a08f3de5591f978a8eb3eedd4890a5ac87d60a3ec0c40665271b52"} Dec 01 16:26:57 crc kubenswrapper[4739]: I1201 16:26:57.894362 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5b0dcc39-d1cc-454a-a403-70d47279541c","Type":"ContainerStarted","Data":"a571b6c958638eafcef691e477afcd94f108497774c3aff88d1c6a7bba8d7d4b"} Dec 01 16:26:57 crc kubenswrapper[4739]: I1201 16:26:57.905967 4739 generic.go:334] "Generic (PLEG): container finished" podID="f27b159a-6738-43d0-8ba1-041cc46170dc" containerID="55c660450dba588c819269833bd82ca38dd329a7dfb18ca09ed45e2db3df0b6e" exitCode=0 Dec 01 16:26:57 crc kubenswrapper[4739]: I1201 16:26:57.906034 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" event={"ID":"f27b159a-6738-43d0-8ba1-041cc46170dc","Type":"ContainerDied","Data":"55c660450dba588c819269833bd82ca38dd329a7dfb18ca09ed45e2db3df0b6e"} Dec 01 16:26:57 crc kubenswrapper[4739]: I1201 16:26:57.906060 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" event={"ID":"f27b159a-6738-43d0-8ba1-041cc46170dc","Type":"ContainerStarted","Data":"88c349ee704701af94c4bcc37e38a7cc28bf381ad476f935365663c48a5eea0a"} Dec 01 16:26:57 crc kubenswrapper[4739]: I1201 16:26:57.910820 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3930d334-34db-4729-9064-140ba9d413f4","Type":"ContainerStarted","Data":"735453f371b7fdd694f79c2f8a1388f5117fb8218566102be137e523cba4a4ee"} Dec 01 16:26:58 crc kubenswrapper[4739]: I1201 16:26:58.929213 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5b0dcc39-d1cc-454a-a403-70d47279541c","Type":"ContainerStarted","Data":"15897500412d62529446a37ce315bde647579396696d1eec233934952bde65a1"} Dec 01 16:26:58 crc kubenswrapper[4739]: I1201 16:26:58.932157 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" event={"ID":"f27b159a-6738-43d0-8ba1-041cc46170dc","Type":"ContainerStarted","Data":"9a04e9a0a58223c74c75daf4069fece5f4767e3219b181775d424582d0d2f7a1"} Dec 01 16:26:58 crc kubenswrapper[4739]: I1201 16:26:58.932201 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:26:58 crc kubenswrapper[4739]: I1201 16:26:58.939377 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"57dfdfae-381a-46ca-921c-2528f1c11d05","Type":"ContainerStarted","Data":"5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab"} Dec 01 16:26:58 crc kubenswrapper[4739]: I1201 16:26:58.939600 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 01 16:26:58 crc kubenswrapper[4739]: I1201 16:26:58.960913 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" podStartSLOduration=2.960895752 podStartE2EDuration="2.960895752s" podCreationTimestamp="2025-12-01 16:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:26:58.94752866 +0000 UTC m=+3120.773274774" watchObservedRunningTime="2025-12-01 16:26:58.960895752 +0000 UTC m=+3120.786641846" Dec 01 16:26:58 crc kubenswrapper[4739]: I1201 16:26:58.986464 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=2.986447062 podStartE2EDuration="2.986447062s" podCreationTimestamp="2025-12-01 16:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:26:58.977101093 +0000 UTC m=+3120.802847207" watchObservedRunningTime="2025-12-01 16:26:58.986447062 +0000 UTC m=+3120.812193156" Dec 01 16:26:59 crc kubenswrapper[4739]: I1201 16:26:59.061449 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 01 16:26:59 crc kubenswrapper[4739]: I1201 16:26:59.949102 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"57dfdfae-381a-46ca-921c-2528f1c11d05","Type":"ContainerStarted","Data":"10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152"} Dec 01 16:26:59 crc kubenswrapper[4739]: I1201 16:26:59.953117 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5b0dcc39-d1cc-454a-a403-70d47279541c","Type":"ContainerStarted","Data":"5df9727619fac02426e2b001a1d2dd68d53964ca1359dcd3f7b83959b9b4be85"} Dec 01 16:26:59 crc kubenswrapper[4739]: I1201 16:26:59.975040 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.15690285 podStartE2EDuration="3.975019311s" podCreationTimestamp="2025-12-01 16:26:56 +0000 UTC" firstStartedPulling="2025-12-01 16:26:57.050135267 +0000 UTC m=+3118.875881361" lastFinishedPulling="2025-12-01 16:26:57.868251728 +0000 UTC m=+3119.693997822" observedRunningTime="2025-12-01 16:26:59.967199869 +0000 UTC m=+3121.792945963" watchObservedRunningTime="2025-12-01 16:26:59.975019311 +0000 UTC m=+3121.800765415" Dec 01 16:27:00 crc kubenswrapper[4739]: I1201 16:27:00.961675 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="57dfdfae-381a-46ca-921c-2528f1c11d05" containerName="manila-api-log" containerID="cri-o://5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab" gracePeriod=30 Dec 01 16:27:00 crc kubenswrapper[4739]: I1201 16:27:00.961773 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="57dfdfae-381a-46ca-921c-2528f1c11d05" containerName="manila-api" containerID="cri-o://10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152" gracePeriod=30 Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.326584 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.418525 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.616579 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.771155 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57dfdfae-381a-46ca-921c-2528f1c11d05-logs\") pod \"57dfdfae-381a-46ca-921c-2528f1c11d05\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.771252 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-scripts\") pod \"57dfdfae-381a-46ca-921c-2528f1c11d05\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.771329 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/57dfdfae-381a-46ca-921c-2528f1c11d05-etc-machine-id\") pod \"57dfdfae-381a-46ca-921c-2528f1c11d05\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.771458 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/57dfdfae-381a-46ca-921c-2528f1c11d05-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "57dfdfae-381a-46ca-921c-2528f1c11d05" (UID: "57dfdfae-381a-46ca-921c-2528f1c11d05"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.771548 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-config-data-custom\") pod \"57dfdfae-381a-46ca-921c-2528f1c11d05\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.771647 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-combined-ca-bundle\") pod \"57dfdfae-381a-46ca-921c-2528f1c11d05\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.771720 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-config-data\") pod \"57dfdfae-381a-46ca-921c-2528f1c11d05\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.771741 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57dfdfae-381a-46ca-921c-2528f1c11d05-logs" (OuterVolumeSpecName: "logs") pod "57dfdfae-381a-46ca-921c-2528f1c11d05" (UID: "57dfdfae-381a-46ca-921c-2528f1c11d05"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.771871 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gk29c\" (UniqueName: \"kubernetes.io/projected/57dfdfae-381a-46ca-921c-2528f1c11d05-kube-api-access-gk29c\") pod \"57dfdfae-381a-46ca-921c-2528f1c11d05\" (UID: \"57dfdfae-381a-46ca-921c-2528f1c11d05\") " Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.772569 4739 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57dfdfae-381a-46ca-921c-2528f1c11d05-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.772591 4739 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/57dfdfae-381a-46ca-921c-2528f1c11d05-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.780749 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "57dfdfae-381a-46ca-921c-2528f1c11d05" (UID: "57dfdfae-381a-46ca-921c-2528f1c11d05"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.782599 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-scripts" (OuterVolumeSpecName: "scripts") pod "57dfdfae-381a-46ca-921c-2528f1c11d05" (UID: "57dfdfae-381a-46ca-921c-2528f1c11d05"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.799073 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57dfdfae-381a-46ca-921c-2528f1c11d05-kube-api-access-gk29c" (OuterVolumeSpecName: "kube-api-access-gk29c") pod "57dfdfae-381a-46ca-921c-2528f1c11d05" (UID: "57dfdfae-381a-46ca-921c-2528f1c11d05"). InnerVolumeSpecName "kube-api-access-gk29c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.822708 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57dfdfae-381a-46ca-921c-2528f1c11d05" (UID: "57dfdfae-381a-46ca-921c-2528f1c11d05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.857990 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-config-data" (OuterVolumeSpecName: "config-data") pod "57dfdfae-381a-46ca-921c-2528f1c11d05" (UID: "57dfdfae-381a-46ca-921c-2528f1c11d05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.875219 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.875249 4739 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.875301 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.875312 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57dfdfae-381a-46ca-921c-2528f1c11d05-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.875324 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gk29c\" (UniqueName: \"kubernetes.io/projected/57dfdfae-381a-46ca-921c-2528f1c11d05-kube-api-access-gk29c\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.973227 4739 generic.go:334] "Generic (PLEG): container finished" podID="57dfdfae-381a-46ca-921c-2528f1c11d05" containerID="10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152" exitCode=0 Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.973279 4739 generic.go:334] "Generic (PLEG): container finished" podID="57dfdfae-381a-46ca-921c-2528f1c11d05" containerID="5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab" exitCode=143 Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.973310 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"57dfdfae-381a-46ca-921c-2528f1c11d05","Type":"ContainerDied","Data":"10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152"} Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.973346 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"57dfdfae-381a-46ca-921c-2528f1c11d05","Type":"ContainerDied","Data":"5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab"} Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.973365 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"57dfdfae-381a-46ca-921c-2528f1c11d05","Type":"ContainerDied","Data":"626b7ebf51a08f3de5591f978a8eb3eedd4890a5ac87d60a3ec0c40665271b52"} Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.973390 4739 scope.go:117] "RemoveContainer" containerID="10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152" Dec 01 16:27:01 crc kubenswrapper[4739]: I1201 16:27:01.973598 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.023496 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.035555 4739 scope.go:117] "RemoveContainer" containerID="5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.039763 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.054301 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 01 16:27:02 crc kubenswrapper[4739]: E1201 16:27:02.054934 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57dfdfae-381a-46ca-921c-2528f1c11d05" containerName="manila-api" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.055015 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="57dfdfae-381a-46ca-921c-2528f1c11d05" containerName="manila-api" Dec 01 16:27:02 crc kubenswrapper[4739]: E1201 16:27:02.055099 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57dfdfae-381a-46ca-921c-2528f1c11d05" containerName="manila-api-log" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.055162 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="57dfdfae-381a-46ca-921c-2528f1c11d05" containerName="manila-api-log" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.055388 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="57dfdfae-381a-46ca-921c-2528f1c11d05" containerName="manila-api" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.055486 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="57dfdfae-381a-46ca-921c-2528f1c11d05" containerName="manila-api-log" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.056580 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.061587 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.061858 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.063833 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.068863 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.185698 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-public-tls-certs\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.185950 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjrhd\" (UniqueName: \"kubernetes.io/projected/ff7067bb-b693-4d14-9a83-af47d37b421a-kube-api-access-tjrhd\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.186024 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.186055 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-scripts\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.186085 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff7067bb-b693-4d14-9a83-af47d37b421a-logs\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.186101 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff7067bb-b693-4d14-9a83-af47d37b421a-etc-machine-id\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.186150 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-internal-tls-certs\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.186260 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-config-data\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.186305 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-config-data-custom\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.288081 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-public-tls-certs\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.288152 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjrhd\" (UniqueName: \"kubernetes.io/projected/ff7067bb-b693-4d14-9a83-af47d37b421a-kube-api-access-tjrhd\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.288248 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.288308 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-scripts\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.288369 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff7067bb-b693-4d14-9a83-af47d37b421a-logs\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.288439 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff7067bb-b693-4d14-9a83-af47d37b421a-etc-machine-id\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.288497 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-internal-tls-certs\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.288613 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-config-data\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.288686 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-config-data-custom\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.288838 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff7067bb-b693-4d14-9a83-af47d37b421a-etc-machine-id\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.289869 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff7067bb-b693-4d14-9a83-af47d37b421a-logs\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.294983 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-scripts\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.295389 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-public-tls-certs\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.295881 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-config-data\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.308140 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.308262 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-internal-tls-certs\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.309091 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff7067bb-b693-4d14-9a83-af47d37b421a-config-data-custom\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.321758 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjrhd\" (UniqueName: \"kubernetes.io/projected/ff7067bb-b693-4d14-9a83-af47d37b421a-kube-api-access-tjrhd\") pod \"manila-api-0\" (UID: \"ff7067bb-b693-4d14-9a83-af47d37b421a\") " pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.448687 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 01 16:27:02 crc kubenswrapper[4739]: I1201 16:27:02.507066 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57dfdfae-381a-46ca-921c-2528f1c11d05" path="/var/lib/kubelet/pods/57dfdfae-381a-46ca-921c-2528f1c11d05/volumes" Dec 01 16:27:03 crc kubenswrapper[4739]: I1201 16:27:03.389092 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:27:03 crc kubenswrapper[4739]: I1201 16:27:03.407926 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-698d484b74-5hhp4" Dec 01 16:27:03 crc kubenswrapper[4739]: I1201 16:27:03.474406 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8675489cb4-9mz7f"] Dec 01 16:27:03 crc kubenswrapper[4739]: I1201 16:27:03.993375 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-8675489cb4-9mz7f" podUID="0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" containerName="horizon-log" containerID="cri-o://82e1bf034e64723a1b9fead779cd64592427fad19be93d92b9185d673b6e38ed" gracePeriod=30 Dec 01 16:27:03 crc kubenswrapper[4739]: I1201 16:27:03.993886 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-8675489cb4-9mz7f" podUID="0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" containerName="horizon" containerID="cri-o://c30f8e2e727a49a2094463f0e8cf3a19335dd473e4173694417d8ca96e055330" gracePeriod=30 Dec 01 16:27:06 crc kubenswrapper[4739]: I1201 16:27:06.368974 4739 scope.go:117] "RemoveContainer" containerID="10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152" Dec 01 16:27:06 crc kubenswrapper[4739]: E1201 16:27:06.369787 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152\": container with ID starting with 10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152 not found: ID does not exist" containerID="10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152" Dec 01 16:27:06 crc kubenswrapper[4739]: I1201 16:27:06.369845 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152"} err="failed to get container status \"10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152\": rpc error: code = NotFound desc = could not find container \"10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152\": container with ID starting with 10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152 not found: ID does not exist" Dec 01 16:27:06 crc kubenswrapper[4739]: I1201 16:27:06.369880 4739 scope.go:117] "RemoveContainer" containerID="5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab" Dec 01 16:27:06 crc kubenswrapper[4739]: E1201 16:27:06.370351 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab\": container with ID starting with 5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab not found: ID does not exist" containerID="5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab" Dec 01 16:27:06 crc kubenswrapper[4739]: I1201 16:27:06.370436 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab"} err="failed to get container status \"5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab\": rpc error: code = NotFound desc = could not find container \"5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab\": container with ID starting with 5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab not found: ID does not exist" Dec 01 16:27:06 crc kubenswrapper[4739]: I1201 16:27:06.370483 4739 scope.go:117] "RemoveContainer" containerID="10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152" Dec 01 16:27:06 crc kubenswrapper[4739]: I1201 16:27:06.370853 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152"} err="failed to get container status \"10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152\": rpc error: code = NotFound desc = could not find container \"10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152\": container with ID starting with 10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152 not found: ID does not exist" Dec 01 16:27:06 crc kubenswrapper[4739]: I1201 16:27:06.370881 4739 scope.go:117] "RemoveContainer" containerID="5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab" Dec 01 16:27:06 crc kubenswrapper[4739]: I1201 16:27:06.371238 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab"} err="failed to get container status \"5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab\": rpc error: code = NotFound desc = could not find container \"5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab\": container with ID starting with 5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab not found: ID does not exist" Dec 01 16:27:06 crc kubenswrapper[4739]: I1201 16:27:06.446722 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:06.852952 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76b5fdb995-pvgg8" Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:06.928483 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-d6twm"] Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:06.928784 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" podUID="fdf75503-6233-48bc-9c4e-66bd2baa2d16" containerName="dnsmasq-dns" containerID="cri-o://f1c382a69ee627f69cfddb0935265aec232c49ebd46b50f28b7152fac2d72c32" gracePeriod=10 Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.024373 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3930d334-34db-4729-9064-140ba9d413f4","Type":"ContainerStarted","Data":"142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd"} Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.056715 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.558362 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.708592 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-ovsdbserver-nb\") pod \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.709022 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qp8qp\" (UniqueName: \"kubernetes.io/projected/fdf75503-6233-48bc-9c4e-66bd2baa2d16-kube-api-access-qp8qp\") pod \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.709058 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-dns-svc\") pod \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.709216 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-config\") pod \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.709349 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-ovsdbserver-sb\") pod \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.709388 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-openstack-edpm-ipam\") pod \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\" (UID: \"fdf75503-6233-48bc-9c4e-66bd2baa2d16\") " Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.725523 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdf75503-6233-48bc-9c4e-66bd2baa2d16-kube-api-access-qp8qp" (OuterVolumeSpecName: "kube-api-access-qp8qp") pod "fdf75503-6233-48bc-9c4e-66bd2baa2d16" (UID: "fdf75503-6233-48bc-9c4e-66bd2baa2d16"). InnerVolumeSpecName "kube-api-access-qp8qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.780328 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "fdf75503-6233-48bc-9c4e-66bd2baa2d16" (UID: "fdf75503-6233-48bc-9c4e-66bd2baa2d16"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.780600 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fdf75503-6233-48bc-9c4e-66bd2baa2d16" (UID: "fdf75503-6233-48bc-9c4e-66bd2baa2d16"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.789831 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fdf75503-6233-48bc-9c4e-66bd2baa2d16" (UID: "fdf75503-6233-48bc-9c4e-66bd2baa2d16"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.794546 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-config" (OuterVolumeSpecName: "config") pod "fdf75503-6233-48bc-9c4e-66bd2baa2d16" (UID: "fdf75503-6233-48bc-9c4e-66bd2baa2d16"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.800107 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fdf75503-6233-48bc-9c4e-66bd2baa2d16" (UID: "fdf75503-6233-48bc-9c4e-66bd2baa2d16"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.812697 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.812728 4739 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.812737 4739 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.812747 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qp8qp\" (UniqueName: \"kubernetes.io/projected/fdf75503-6233-48bc-9c4e-66bd2baa2d16-kube-api-access-qp8qp\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.812756 4739 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:07 crc kubenswrapper[4739]: I1201 16:27:07.812765 4739 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdf75503-6233-48bc-9c4e-66bd2baa2d16-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.045776 4739 generic.go:334] "Generic (PLEG): container finished" podID="fdf75503-6233-48bc-9c4e-66bd2baa2d16" containerID="f1c382a69ee627f69cfddb0935265aec232c49ebd46b50f28b7152fac2d72c32" exitCode=0 Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.045825 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" event={"ID":"fdf75503-6233-48bc-9c4e-66bd2baa2d16","Type":"ContainerDied","Data":"f1c382a69ee627f69cfddb0935265aec232c49ebd46b50f28b7152fac2d72c32"} Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.045854 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.046093 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864d5fc68c-d6twm" event={"ID":"fdf75503-6233-48bc-9c4e-66bd2baa2d16","Type":"ContainerDied","Data":"68ab754fb6d0cff8bc76d145125c542f428b6972d32c877291877c61a308af82"} Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.046130 4739 scope.go:117] "RemoveContainer" containerID="f1c382a69ee627f69cfddb0935265aec232c49ebd46b50f28b7152fac2d72c32" Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.053033 4739 generic.go:334] "Generic (PLEG): container finished" podID="0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" containerID="c30f8e2e727a49a2094463f0e8cf3a19335dd473e4173694417d8ca96e055330" exitCode=0 Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.053105 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8675489cb4-9mz7f" event={"ID":"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7","Type":"ContainerDied","Data":"c30f8e2e727a49a2094463f0e8cf3a19335dd473e4173694417d8ca96e055330"} Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.055144 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3930d334-34db-4729-9064-140ba9d413f4","Type":"ContainerStarted","Data":"3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298"} Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.065670 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"ff7067bb-b693-4d14-9a83-af47d37b421a","Type":"ContainerStarted","Data":"c070fb9214f5ab8f173caa7242f55f1b915a10cf067886a19eb92d6049d5bc5d"} Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.065710 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"ff7067bb-b693-4d14-9a83-af47d37b421a","Type":"ContainerStarted","Data":"f1bc66ca259fdba07eafd118931b61d129dd34c0f4e10ed38f4ab95d6d0094e5"} Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.084010 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.94432039 podStartE2EDuration="12.083991794s" podCreationTimestamp="2025-12-01 16:26:56 +0000 UTC" firstStartedPulling="2025-12-01 16:26:57.269091623 +0000 UTC m=+3119.094837717" lastFinishedPulling="2025-12-01 16:27:06.408763017 +0000 UTC m=+3128.234509121" observedRunningTime="2025-12-01 16:27:08.077602387 +0000 UTC m=+3129.903348501" watchObservedRunningTime="2025-12-01 16:27:08.083991794 +0000 UTC m=+3129.909737918" Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.112340 4739 scope.go:117] "RemoveContainer" containerID="fc24315c126dc12e60ab5ed5df5094ff8f67b44a60e97ec1864ecbaebec58eb5" Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.120375 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-d6twm"] Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.128928 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-864d5fc68c-d6twm"] Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.144188 4739 scope.go:117] "RemoveContainer" containerID="f1c382a69ee627f69cfddb0935265aec232c49ebd46b50f28b7152fac2d72c32" Dec 01 16:27:08 crc kubenswrapper[4739]: E1201 16:27:08.144570 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1c382a69ee627f69cfddb0935265aec232c49ebd46b50f28b7152fac2d72c32\": container with ID starting with f1c382a69ee627f69cfddb0935265aec232c49ebd46b50f28b7152fac2d72c32 not found: ID does not exist" containerID="f1c382a69ee627f69cfddb0935265aec232c49ebd46b50f28b7152fac2d72c32" Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.144600 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1c382a69ee627f69cfddb0935265aec232c49ebd46b50f28b7152fac2d72c32"} err="failed to get container status \"f1c382a69ee627f69cfddb0935265aec232c49ebd46b50f28b7152fac2d72c32\": rpc error: code = NotFound desc = could not find container \"f1c382a69ee627f69cfddb0935265aec232c49ebd46b50f28b7152fac2d72c32\": container with ID starting with f1c382a69ee627f69cfddb0935265aec232c49ebd46b50f28b7152fac2d72c32 not found: ID does not exist" Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.144622 4739 scope.go:117] "RemoveContainer" containerID="fc24315c126dc12e60ab5ed5df5094ff8f67b44a60e97ec1864ecbaebec58eb5" Dec 01 16:27:08 crc kubenswrapper[4739]: E1201 16:27:08.145058 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc24315c126dc12e60ab5ed5df5094ff8f67b44a60e97ec1864ecbaebec58eb5\": container with ID starting with fc24315c126dc12e60ab5ed5df5094ff8f67b44a60e97ec1864ecbaebec58eb5 not found: ID does not exist" containerID="fc24315c126dc12e60ab5ed5df5094ff8f67b44a60e97ec1864ecbaebec58eb5" Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.145082 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc24315c126dc12e60ab5ed5df5094ff8f67b44a60e97ec1864ecbaebec58eb5"} err="failed to get container status \"fc24315c126dc12e60ab5ed5df5094ff8f67b44a60e97ec1864ecbaebec58eb5\": rpc error: code = NotFound desc = could not find container \"fc24315c126dc12e60ab5ed5df5094ff8f67b44a60e97ec1864ecbaebec58eb5\": container with ID starting with fc24315c126dc12e60ab5ed5df5094ff8f67b44a60e97ec1864ecbaebec58eb5 not found: ID does not exist" Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.487049 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:27:08 crc kubenswrapper[4739]: E1201 16:27:08.487332 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.488682 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdf75503-6233-48bc-9c4e-66bd2baa2d16" path="/var/lib/kubelet/pods/fdf75503-6233-48bc-9c4e-66bd2baa2d16/volumes" Dec 01 16:27:08 crc kubenswrapper[4739]: I1201 16:27:08.718590 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-8675489cb4-9mz7f" podUID="0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.240:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.240:8443: connect: connection refused" Dec 01 16:27:08 crc kubenswrapper[4739]: W1201 16:27:08.736484 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57dfdfae_381a_46ca_921c_2528f1c11d05.slice/crio-10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152.scope WatchSource:0}: Error finding container 10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152: Status 404 returned error can't find the container with id 10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152 Dec 01 16:27:08 crc kubenswrapper[4739]: E1201 16:27:08.963136 4739 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57dfdfae_381a_46ca_921c_2528f1c11d05.slice/crio-5863235b366c1fc89219fe6c06be204c725f68cb4b1ffc9df03964f4ae5b2aab.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdf75503_6233_48bc_9c4e_66bd2baa2d16.slice/crio-68ab754fb6d0cff8bc76d145125c542f428b6972d32c877291877c61a308af82\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0f951fa_e09f_45ec_8110_9a7e9a1f6952.slice/crio-conmon-101e1b3a3b5f4f8110f6505354518dcc71eaa944706c9e6a66750759257ee253.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdf75503_6233_48bc_9c4e_66bd2baa2d16.slice/crio-f1c382a69ee627f69cfddb0935265aec232c49ebd46b50f28b7152fac2d72c32.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0f951fa_e09f_45ec_8110_9a7e9a1f6952.slice/crio-101e1b3a3b5f4f8110f6505354518dcc71eaa944706c9e6a66750759257ee253.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b0dcc39_d1cc_454a_a403_70d47279541c.slice/crio-conmon-15897500412d62529446a37ce315bde647579396696d1eec233934952bde65a1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9e319ab_2c20_4c46_910d_980a0221a139.slice/crio-4d5f9636692ba65ab6b56afe3e47785bfa8a57ec307a0eeb7cb1d1e1eb901780.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9e319ab_2c20_4c46_910d_980a0221a139.slice/crio-conmon-f467346191e267478cd2ad6de53a7393c4b56dbedc0348a05186340f1d422fd5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0dd0df7f_ae84_4ae8_8e0d_bdf5941b4ce7.slice/crio-conmon-c30f8e2e727a49a2094463f0e8cf3a19335dd473e4173694417d8ca96e055330.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57dfdfae_381a_46ca_921c_2528f1c11d05.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0f951fa_e09f_45ec_8110_9a7e9a1f6952.slice/crio-24d2e3b604ad8720e2fab2b7f8d6f358e4b3560908a5583d4071035547398cff.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57dfdfae_381a_46ca_921c_2528f1c11d05.slice/crio-626b7ebf51a08f3de5591f978a8eb3eedd4890a5ac87d60a3ec0c40665271b52\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9e319ab_2c20_4c46_910d_980a0221a139.slice/crio-f467346191e267478cd2ad6de53a7393c4b56dbedc0348a05186340f1d422fd5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdf75503_6233_48bc_9c4e_66bd2baa2d16.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0dd0df7f_ae84_4ae8_8e0d_bdf5941b4ce7.slice/crio-c30f8e2e727a49a2094463f0e8cf3a19335dd473e4173694417d8ca96e055330.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdf75503_6233_48bc_9c4e_66bd2baa2d16.slice/crio-conmon-f1c382a69ee627f69cfddb0935265aec232c49ebd46b50f28b7152fac2d72c32.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57dfdfae_381a_46ca_921c_2528f1c11d05.slice/crio-conmon-10e9b5f066370bbce5307a1d9bb65268134e8c7bb19b3be1fca9be4acdd44152.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0f951fa_e09f_45ec_8110_9a7e9a1f6952.slice/crio-conmon-24d2e3b604ad8720e2fab2b7f8d6f358e4b3560908a5583d4071035547398cff.scope\": RecentStats: unable to find data in memory cache]" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.082396 4739 generic.go:334] "Generic (PLEG): container finished" podID="e9e319ab-2c20-4c46-910d-980a0221a139" containerID="4d5f9636692ba65ab6b56afe3e47785bfa8a57ec307a0eeb7cb1d1e1eb901780" exitCode=137 Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.082798 4739 generic.go:334] "Generic (PLEG): container finished" podID="e9e319ab-2c20-4c46-910d-980a0221a139" containerID="f467346191e267478cd2ad6de53a7393c4b56dbedc0348a05186340f1d422fd5" exitCode=137 Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.082840 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f698fbd6f-nmnlm" event={"ID":"e9e319ab-2c20-4c46-910d-980a0221a139","Type":"ContainerDied","Data":"4d5f9636692ba65ab6b56afe3e47785bfa8a57ec307a0eeb7cb1d1e1eb901780"} Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.082863 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f698fbd6f-nmnlm" event={"ID":"e9e319ab-2c20-4c46-910d-980a0221a139","Type":"ContainerDied","Data":"f467346191e267478cd2ad6de53a7393c4b56dbedc0348a05186340f1d422fd5"} Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.087166 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"ff7067bb-b693-4d14-9a83-af47d37b421a","Type":"ContainerStarted","Data":"8f78481ef5194c62b3e41928df128cc082da43c54f7c24cc3228b337096ea383"} Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.088366 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.098528 4739 generic.go:334] "Generic (PLEG): container finished" podID="a0f951fa-e09f-45ec-8110-9a7e9a1f6952" containerID="101e1b3a3b5f4f8110f6505354518dcc71eaa944706c9e6a66750759257ee253" exitCode=137 Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.098575 4739 generic.go:334] "Generic (PLEG): container finished" podID="a0f951fa-e09f-45ec-8110-9a7e9a1f6952" containerID="24d2e3b604ad8720e2fab2b7f8d6f358e4b3560908a5583d4071035547398cff" exitCode=137 Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.099557 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f84667477-69xkr" event={"ID":"a0f951fa-e09f-45ec-8110-9a7e9a1f6952","Type":"ContainerDied","Data":"101e1b3a3b5f4f8110f6505354518dcc71eaa944706c9e6a66750759257ee253"} Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.099596 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f84667477-69xkr" event={"ID":"a0f951fa-e09f-45ec-8110-9a7e9a1f6952","Type":"ContainerDied","Data":"24d2e3b604ad8720e2fab2b7f8d6f358e4b3560908a5583d4071035547398cff"} Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.131131 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=7.131096111 podStartE2EDuration="7.131096111s" podCreationTimestamp="2025-12-01 16:27:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:27:09.105755519 +0000 UTC m=+3130.931501613" watchObservedRunningTime="2025-12-01 16:27:09.131096111 +0000 UTC m=+3130.956842205" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.265829 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.274457 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.446921 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e9e319ab-2c20-4c46-910d-980a0221a139-horizon-secret-key\") pod \"e9e319ab-2c20-4c46-910d-980a0221a139\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.447254 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9e319ab-2c20-4c46-910d-980a0221a139-logs\") pod \"e9e319ab-2c20-4c46-910d-980a0221a139\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.447331 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-horizon-secret-key\") pod \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.447396 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-config-data\") pod \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.447463 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-scripts\") pod \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.447497 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e9e319ab-2c20-4c46-910d-980a0221a139-config-data\") pod \"e9e319ab-2c20-4c46-910d-980a0221a139\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.447551 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvwxn\" (UniqueName: \"kubernetes.io/projected/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-kube-api-access-qvwxn\") pod \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.447593 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9e319ab-2c20-4c46-910d-980a0221a139-scripts\") pod \"e9e319ab-2c20-4c46-910d-980a0221a139\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.447647 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z85v9\" (UniqueName: \"kubernetes.io/projected/e9e319ab-2c20-4c46-910d-980a0221a139-kube-api-access-z85v9\") pod \"e9e319ab-2c20-4c46-910d-980a0221a139\" (UID: \"e9e319ab-2c20-4c46-910d-980a0221a139\") " Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.447679 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-logs\") pod \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\" (UID: \"a0f951fa-e09f-45ec-8110-9a7e9a1f6952\") " Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.449731 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-logs" (OuterVolumeSpecName: "logs") pod "a0f951fa-e09f-45ec-8110-9a7e9a1f6952" (UID: "a0f951fa-e09f-45ec-8110-9a7e9a1f6952"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.449983 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9e319ab-2c20-4c46-910d-980a0221a139-logs" (OuterVolumeSpecName: "logs") pod "e9e319ab-2c20-4c46-910d-980a0221a139" (UID: "e9e319ab-2c20-4c46-910d-980a0221a139"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.454279 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-kube-api-access-qvwxn" (OuterVolumeSpecName: "kube-api-access-qvwxn") pod "a0f951fa-e09f-45ec-8110-9a7e9a1f6952" (UID: "a0f951fa-e09f-45ec-8110-9a7e9a1f6952"). InnerVolumeSpecName "kube-api-access-qvwxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.456827 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9e319ab-2c20-4c46-910d-980a0221a139-kube-api-access-z85v9" (OuterVolumeSpecName: "kube-api-access-z85v9") pod "e9e319ab-2c20-4c46-910d-980a0221a139" (UID: "e9e319ab-2c20-4c46-910d-980a0221a139"). InnerVolumeSpecName "kube-api-access-z85v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.458481 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9e319ab-2c20-4c46-910d-980a0221a139-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e9e319ab-2c20-4c46-910d-980a0221a139" (UID: "e9e319ab-2c20-4c46-910d-980a0221a139"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.461792 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a0f951fa-e09f-45ec-8110-9a7e9a1f6952" (UID: "a0f951fa-e09f-45ec-8110-9a7e9a1f6952"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.486860 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-config-data" (OuterVolumeSpecName: "config-data") pod "a0f951fa-e09f-45ec-8110-9a7e9a1f6952" (UID: "a0f951fa-e09f-45ec-8110-9a7e9a1f6952"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.502950 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9e319ab-2c20-4c46-910d-980a0221a139-scripts" (OuterVolumeSpecName: "scripts") pod "e9e319ab-2c20-4c46-910d-980a0221a139" (UID: "e9e319ab-2c20-4c46-910d-980a0221a139"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.507256 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9e319ab-2c20-4c46-910d-980a0221a139-config-data" (OuterVolumeSpecName: "config-data") pod "e9e319ab-2c20-4c46-910d-980a0221a139" (UID: "e9e319ab-2c20-4c46-910d-980a0221a139"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.525589 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-scripts" (OuterVolumeSpecName: "scripts") pod "a0f951fa-e09f-45ec-8110-9a7e9a1f6952" (UID: "a0f951fa-e09f-45ec-8110-9a7e9a1f6952"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.551071 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvwxn\" (UniqueName: \"kubernetes.io/projected/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-kube-api-access-qvwxn\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.551114 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9e319ab-2c20-4c46-910d-980a0221a139-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.551127 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z85v9\" (UniqueName: \"kubernetes.io/projected/e9e319ab-2c20-4c46-910d-980a0221a139-kube-api-access-z85v9\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.551139 4739 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.551153 4739 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e9e319ab-2c20-4c46-910d-980a0221a139-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.551166 4739 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9e319ab-2c20-4c46-910d-980a0221a139-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.551178 4739 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.551189 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.551202 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0f951fa-e09f-45ec-8110-9a7e9a1f6952-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.551211 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e9e319ab-2c20-4c46-910d-980a0221a139-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.647664 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.647922 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="ceilometer-central-agent" containerID="cri-o://5afacc5853ae439e6e32109942743089d84b1baeba5b9df62165dbf87ab3c2f4" gracePeriod=30 Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.648032 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="proxy-httpd" containerID="cri-o://710683be135103ccee71ba86e78c898cd9863ecfe958ba0b6783e508b9b83998" gracePeriod=30 Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.648070 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="sg-core" containerID="cri-o://5954d26bc5a08aa1686874d52e97a553d0c0a97f6083ae9eda383c70da6683f6" gracePeriod=30 Dec 01 16:27:09 crc kubenswrapper[4739]: I1201 16:27:09.648102 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="ceilometer-notification-agent" containerID="cri-o://7cea978622b07b42b7ec40beeec9a14232afcc1ad2b4e521453f7d8fb9d8eb12" gracePeriod=30 Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.109914 4739 generic.go:334] "Generic (PLEG): container finished" podID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerID="710683be135103ccee71ba86e78c898cd9863ecfe958ba0b6783e508b9b83998" exitCode=0 Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.110223 4739 generic.go:334] "Generic (PLEG): container finished" podID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerID="5954d26bc5a08aa1686874d52e97a553d0c0a97f6083ae9eda383c70da6683f6" exitCode=2 Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.110232 4739 generic.go:334] "Generic (PLEG): container finished" podID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerID="5afacc5853ae439e6e32109942743089d84b1baeba5b9df62165dbf87ab3c2f4" exitCode=0 Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.109981 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37c0dad8-d954-4bae-8654-68085a9eb6d3","Type":"ContainerDied","Data":"710683be135103ccee71ba86e78c898cd9863ecfe958ba0b6783e508b9b83998"} Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.110284 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37c0dad8-d954-4bae-8654-68085a9eb6d3","Type":"ContainerDied","Data":"5954d26bc5a08aa1686874d52e97a553d0c0a97f6083ae9eda383c70da6683f6"} Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.110297 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37c0dad8-d954-4bae-8654-68085a9eb6d3","Type":"ContainerDied","Data":"5afacc5853ae439e6e32109942743089d84b1baeba5b9df62165dbf87ab3c2f4"} Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.112517 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f698fbd6f-nmnlm" event={"ID":"e9e319ab-2c20-4c46-910d-980a0221a139","Type":"ContainerDied","Data":"c93ca7e3a8547c86666650f9320fe20376fc31c4992f7b624240ba00fe24ae9c"} Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.112576 4739 scope.go:117] "RemoveContainer" containerID="4d5f9636692ba65ab6b56afe3e47785bfa8a57ec307a0eeb7cb1d1e1eb901780" Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.112713 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f698fbd6f-nmnlm" Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.115367 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f84667477-69xkr" event={"ID":"a0f951fa-e09f-45ec-8110-9a7e9a1f6952","Type":"ContainerDied","Data":"c5f81574844509023f3e28a2b21fa8368777bfc5422461fe9aad75bde2399ffd"} Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.115491 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f84667477-69xkr" Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.183760 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f698fbd6f-nmnlm"] Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.192461 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7f698fbd6f-nmnlm"] Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.201047 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f84667477-69xkr"] Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.216454 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7f84667477-69xkr"] Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.265635 4739 scope.go:117] "RemoveContainer" containerID="f467346191e267478cd2ad6de53a7393c4b56dbedc0348a05186340f1d422fd5" Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.287878 4739 scope.go:117] "RemoveContainer" containerID="101e1b3a3b5f4f8110f6505354518dcc71eaa944706c9e6a66750759257ee253" Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.470063 4739 scope.go:117] "RemoveContainer" containerID="24d2e3b604ad8720e2fab2b7f8d6f358e4b3560908a5583d4071035547398cff" Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.490986 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0f951fa-e09f-45ec-8110-9a7e9a1f6952" path="/var/lib/kubelet/pods/a0f951fa-e09f-45ec-8110-9a7e9a1f6952/volumes" Dec 01 16:27:10 crc kubenswrapper[4739]: I1201 16:27:10.491727 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9e319ab-2c20-4c46-910d-980a0221a139" path="/var/lib/kubelet/pods/e9e319ab-2c20-4c46-910d-980a0221a139/volumes" Dec 01 16:27:11 crc kubenswrapper[4739]: I1201 16:27:11.852969 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.001916 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37c0dad8-d954-4bae-8654-68085a9eb6d3-run-httpd\") pod \"37c0dad8-d954-4bae-8654-68085a9eb6d3\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.001986 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-sg-core-conf-yaml\") pod \"37c0dad8-d954-4bae-8654-68085a9eb6d3\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.002009 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-combined-ca-bundle\") pod \"37c0dad8-d954-4bae-8654-68085a9eb6d3\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.002074 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-config-data\") pod \"37c0dad8-d954-4bae-8654-68085a9eb6d3\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.002136 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-ceilometer-tls-certs\") pod \"37c0dad8-d954-4bae-8654-68085a9eb6d3\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.002157 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-scripts\") pod \"37c0dad8-d954-4bae-8654-68085a9eb6d3\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.002180 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk4h4\" (UniqueName: \"kubernetes.io/projected/37c0dad8-d954-4bae-8654-68085a9eb6d3-kube-api-access-dk4h4\") pod \"37c0dad8-d954-4bae-8654-68085a9eb6d3\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.002256 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37c0dad8-d954-4bae-8654-68085a9eb6d3-log-httpd\") pod \"37c0dad8-d954-4bae-8654-68085a9eb6d3\" (UID: \"37c0dad8-d954-4bae-8654-68085a9eb6d3\") " Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.003308 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37c0dad8-d954-4bae-8654-68085a9eb6d3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "37c0dad8-d954-4bae-8654-68085a9eb6d3" (UID: "37c0dad8-d954-4bae-8654-68085a9eb6d3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.004625 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37c0dad8-d954-4bae-8654-68085a9eb6d3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "37c0dad8-d954-4bae-8654-68085a9eb6d3" (UID: "37c0dad8-d954-4bae-8654-68085a9eb6d3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.007798 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37c0dad8-d954-4bae-8654-68085a9eb6d3-kube-api-access-dk4h4" (OuterVolumeSpecName: "kube-api-access-dk4h4") pod "37c0dad8-d954-4bae-8654-68085a9eb6d3" (UID: "37c0dad8-d954-4bae-8654-68085a9eb6d3"). InnerVolumeSpecName "kube-api-access-dk4h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.024998 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-scripts" (OuterVolumeSpecName: "scripts") pod "37c0dad8-d954-4bae-8654-68085a9eb6d3" (UID: "37c0dad8-d954-4bae-8654-68085a9eb6d3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.035851 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "37c0dad8-d954-4bae-8654-68085a9eb6d3" (UID: "37c0dad8-d954-4bae-8654-68085a9eb6d3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.064235 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "37c0dad8-d954-4bae-8654-68085a9eb6d3" (UID: "37c0dad8-d954-4bae-8654-68085a9eb6d3"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.089190 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37c0dad8-d954-4bae-8654-68085a9eb6d3" (UID: "37c0dad8-d954-4bae-8654-68085a9eb6d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.107395 4739 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37c0dad8-d954-4bae-8654-68085a9eb6d3-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.107511 4739 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.107577 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.107597 4739 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.107660 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.107689 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk4h4\" (UniqueName: \"kubernetes.io/projected/37c0dad8-d954-4bae-8654-68085a9eb6d3-kube-api-access-dk4h4\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.107767 4739 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37c0dad8-d954-4bae-8654-68085a9eb6d3-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.131296 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-config-data" (OuterVolumeSpecName: "config-data") pod "37c0dad8-d954-4bae-8654-68085a9eb6d3" (UID: "37c0dad8-d954-4bae-8654-68085a9eb6d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.143568 4739 generic.go:334] "Generic (PLEG): container finished" podID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerID="7cea978622b07b42b7ec40beeec9a14232afcc1ad2b4e521453f7d8fb9d8eb12" exitCode=0 Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.143613 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37c0dad8-d954-4bae-8654-68085a9eb6d3","Type":"ContainerDied","Data":"7cea978622b07b42b7ec40beeec9a14232afcc1ad2b4e521453f7d8fb9d8eb12"} Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.143644 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37c0dad8-d954-4bae-8654-68085a9eb6d3","Type":"ContainerDied","Data":"345c8dbccec64e8b4bf7ed6c15474b8ebf928d6ef8a6d2ec0cdefe8c88ff3e65"} Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.143663 4739 scope.go:117] "RemoveContainer" containerID="710683be135103ccee71ba86e78c898cd9863ecfe958ba0b6783e508b9b83998" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.143698 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.178931 4739 scope.go:117] "RemoveContainer" containerID="5954d26bc5a08aa1686874d52e97a553d0c0a97f6083ae9eda383c70da6683f6" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.194512 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.204878 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.210598 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37c0dad8-d954-4bae-8654-68085a9eb6d3-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.219388 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 16:27:12 crc kubenswrapper[4739]: E1201 16:27:12.219763 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f951fa-e09f-45ec-8110-9a7e9a1f6952" containerName="horizon-log" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.219774 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f951fa-e09f-45ec-8110-9a7e9a1f6952" containerName="horizon-log" Dec 01 16:27:12 crc kubenswrapper[4739]: E1201 16:27:12.219782 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9e319ab-2c20-4c46-910d-980a0221a139" containerName="horizon" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.219788 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9e319ab-2c20-4c46-910d-980a0221a139" containerName="horizon" Dec 01 16:27:12 crc kubenswrapper[4739]: E1201 16:27:12.219798 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="sg-core" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.219805 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="sg-core" Dec 01 16:27:12 crc kubenswrapper[4739]: E1201 16:27:12.219820 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f951fa-e09f-45ec-8110-9a7e9a1f6952" containerName="horizon" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.219827 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f951fa-e09f-45ec-8110-9a7e9a1f6952" containerName="horizon" Dec 01 16:27:12 crc kubenswrapper[4739]: E1201 16:27:12.219840 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdf75503-6233-48bc-9c4e-66bd2baa2d16" containerName="init" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.219861 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdf75503-6233-48bc-9c4e-66bd2baa2d16" containerName="init" Dec 01 16:27:12 crc kubenswrapper[4739]: E1201 16:27:12.219873 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="proxy-httpd" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.219878 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="proxy-httpd" Dec 01 16:27:12 crc kubenswrapper[4739]: E1201 16:27:12.219898 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9e319ab-2c20-4c46-910d-980a0221a139" containerName="horizon-log" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.219903 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9e319ab-2c20-4c46-910d-980a0221a139" containerName="horizon-log" Dec 01 16:27:12 crc kubenswrapper[4739]: E1201 16:27:12.219913 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdf75503-6233-48bc-9c4e-66bd2baa2d16" containerName="dnsmasq-dns" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.219919 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdf75503-6233-48bc-9c4e-66bd2baa2d16" containerName="dnsmasq-dns" Dec 01 16:27:12 crc kubenswrapper[4739]: E1201 16:27:12.219930 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="ceilometer-notification-agent" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.219935 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="ceilometer-notification-agent" Dec 01 16:27:12 crc kubenswrapper[4739]: E1201 16:27:12.219954 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="ceilometer-central-agent" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.219960 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="ceilometer-central-agent" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.220117 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0f951fa-e09f-45ec-8110-9a7e9a1f6952" containerName="horizon-log" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.220130 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="sg-core" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.220139 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="ceilometer-central-agent" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.220182 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="proxy-httpd" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.220193 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0f951fa-e09f-45ec-8110-9a7e9a1f6952" containerName="horizon" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.220205 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9e319ab-2c20-4c46-910d-980a0221a139" containerName="horizon-log" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.220214 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" containerName="ceilometer-notification-agent" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.220226 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdf75503-6233-48bc-9c4e-66bd2baa2d16" containerName="dnsmasq-dns" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.220234 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9e319ab-2c20-4c46-910d-980a0221a139" containerName="horizon" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.222387 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.229746 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.230038 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.232595 4739 scope.go:117] "RemoveContainer" containerID="7cea978622b07b42b7ec40beeec9a14232afcc1ad2b4e521453f7d8fb9d8eb12" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.232853 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.241032 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.303347 4739 scope.go:117] "RemoveContainer" containerID="5afacc5853ae439e6e32109942743089d84b1baeba5b9df62165dbf87ab3c2f4" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.320476 4739 scope.go:117] "RemoveContainer" containerID="710683be135103ccee71ba86e78c898cd9863ecfe958ba0b6783e508b9b83998" Dec 01 16:27:12 crc kubenswrapper[4739]: E1201 16:27:12.320983 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"710683be135103ccee71ba86e78c898cd9863ecfe958ba0b6783e508b9b83998\": container with ID starting with 710683be135103ccee71ba86e78c898cd9863ecfe958ba0b6783e508b9b83998 not found: ID does not exist" containerID="710683be135103ccee71ba86e78c898cd9863ecfe958ba0b6783e508b9b83998" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.321027 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"710683be135103ccee71ba86e78c898cd9863ecfe958ba0b6783e508b9b83998"} err="failed to get container status \"710683be135103ccee71ba86e78c898cd9863ecfe958ba0b6783e508b9b83998\": rpc error: code = NotFound desc = could not find container \"710683be135103ccee71ba86e78c898cd9863ecfe958ba0b6783e508b9b83998\": container with ID starting with 710683be135103ccee71ba86e78c898cd9863ecfe958ba0b6783e508b9b83998 not found: ID does not exist" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.321055 4739 scope.go:117] "RemoveContainer" containerID="5954d26bc5a08aa1686874d52e97a553d0c0a97f6083ae9eda383c70da6683f6" Dec 01 16:27:12 crc kubenswrapper[4739]: E1201 16:27:12.321482 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5954d26bc5a08aa1686874d52e97a553d0c0a97f6083ae9eda383c70da6683f6\": container with ID starting with 5954d26bc5a08aa1686874d52e97a553d0c0a97f6083ae9eda383c70da6683f6 not found: ID does not exist" containerID="5954d26bc5a08aa1686874d52e97a553d0c0a97f6083ae9eda383c70da6683f6" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.321569 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5954d26bc5a08aa1686874d52e97a553d0c0a97f6083ae9eda383c70da6683f6"} err="failed to get container status \"5954d26bc5a08aa1686874d52e97a553d0c0a97f6083ae9eda383c70da6683f6\": rpc error: code = NotFound desc = could not find container \"5954d26bc5a08aa1686874d52e97a553d0c0a97f6083ae9eda383c70da6683f6\": container with ID starting with 5954d26bc5a08aa1686874d52e97a553d0c0a97f6083ae9eda383c70da6683f6 not found: ID does not exist" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.321637 4739 scope.go:117] "RemoveContainer" containerID="7cea978622b07b42b7ec40beeec9a14232afcc1ad2b4e521453f7d8fb9d8eb12" Dec 01 16:27:12 crc kubenswrapper[4739]: E1201 16:27:12.322023 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cea978622b07b42b7ec40beeec9a14232afcc1ad2b4e521453f7d8fb9d8eb12\": container with ID starting with 7cea978622b07b42b7ec40beeec9a14232afcc1ad2b4e521453f7d8fb9d8eb12 not found: ID does not exist" containerID="7cea978622b07b42b7ec40beeec9a14232afcc1ad2b4e521453f7d8fb9d8eb12" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.322064 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cea978622b07b42b7ec40beeec9a14232afcc1ad2b4e521453f7d8fb9d8eb12"} err="failed to get container status \"7cea978622b07b42b7ec40beeec9a14232afcc1ad2b4e521453f7d8fb9d8eb12\": rpc error: code = NotFound desc = could not find container \"7cea978622b07b42b7ec40beeec9a14232afcc1ad2b4e521453f7d8fb9d8eb12\": container with ID starting with 7cea978622b07b42b7ec40beeec9a14232afcc1ad2b4e521453f7d8fb9d8eb12 not found: ID does not exist" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.322085 4739 scope.go:117] "RemoveContainer" containerID="5afacc5853ae439e6e32109942743089d84b1baeba5b9df62165dbf87ab3c2f4" Dec 01 16:27:12 crc kubenswrapper[4739]: E1201 16:27:12.322344 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5afacc5853ae439e6e32109942743089d84b1baeba5b9df62165dbf87ab3c2f4\": container with ID starting with 5afacc5853ae439e6e32109942743089d84b1baeba5b9df62165dbf87ab3c2f4 not found: ID does not exist" containerID="5afacc5853ae439e6e32109942743089d84b1baeba5b9df62165dbf87ab3c2f4" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.322370 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5afacc5853ae439e6e32109942743089d84b1baeba5b9df62165dbf87ab3c2f4"} err="failed to get container status \"5afacc5853ae439e6e32109942743089d84b1baeba5b9df62165dbf87ab3c2f4\": rpc error: code = NotFound desc = could not find container \"5afacc5853ae439e6e32109942743089d84b1baeba5b9df62165dbf87ab3c2f4\": container with ID starting with 5afacc5853ae439e6e32109942743089d84b1baeba5b9df62165dbf87ab3c2f4 not found: ID does not exist" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.414345 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.414620 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-log-httpd\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.414780 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-run-httpd\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.414884 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-scripts\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.414981 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xnvt\" (UniqueName: \"kubernetes.io/projected/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-kube-api-access-2xnvt\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.415087 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.415165 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.415249 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-config-data\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.488954 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37c0dad8-d954-4bae-8654-68085a9eb6d3" path="/var/lib/kubelet/pods/37c0dad8-d954-4bae-8654-68085a9eb6d3/volumes" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.525190 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.525254 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-log-httpd\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.525338 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-run-httpd\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.525371 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-scripts\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.525412 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xnvt\" (UniqueName: \"kubernetes.io/projected/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-kube-api-access-2xnvt\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.525461 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.525490 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.525553 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-config-data\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.526746 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-log-httpd\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.526924 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-run-httpd\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.531058 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-scripts\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.531046 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.531464 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.538213 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-config-data\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.541139 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.542720 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xnvt\" (UniqueName: \"kubernetes.io/projected/7bacd19a-d9e6-4894-bc04-c11b7c5f3c85-kube-api-access-2xnvt\") pod \"ceilometer-0\" (UID: \"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85\") " pod="openstack/ceilometer-0" Dec 01 16:27:12 crc kubenswrapper[4739]: I1201 16:27:12.543651 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 16:27:13 crc kubenswrapper[4739]: I1201 16:27:13.085123 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 16:27:13 crc kubenswrapper[4739]: W1201 16:27:13.087162 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bacd19a_d9e6_4894_bc04_c11b7c5f3c85.slice/crio-7a79bf3bf039651c4cc509971abe8c9e1c535c99d6d4f81efef00795c6978fdd WatchSource:0}: Error finding container 7a79bf3bf039651c4cc509971abe8c9e1c535c99d6d4f81efef00795c6978fdd: Status 404 returned error can't find the container with id 7a79bf3bf039651c4cc509971abe8c9e1c535c99d6d4f81efef00795c6978fdd Dec 01 16:27:13 crc kubenswrapper[4739]: I1201 16:27:13.155361 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85","Type":"ContainerStarted","Data":"7a79bf3bf039651c4cc509971abe8c9e1c535c99d6d4f81efef00795c6978fdd"} Dec 01 16:27:14 crc kubenswrapper[4739]: I1201 16:27:14.170236 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85","Type":"ContainerStarted","Data":"a49c0e5b004557542cddbbedeceb25964ad2edf5e8b698bca71ba6e2fe83ba4d"} Dec 01 16:27:16 crc kubenswrapper[4739]: I1201 16:27:16.196944 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85","Type":"ContainerStarted","Data":"a8dbdcb3bbaf4b4345d41a730fce53cb487511659bff58a92f7af185b32ba23e"} Dec 01 16:27:16 crc kubenswrapper[4739]: I1201 16:27:16.558442 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 01 16:27:17 crc kubenswrapper[4739]: I1201 16:27:17.221946 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85","Type":"ContainerStarted","Data":"9141aff68ef11b3525a8b978a713518e6ca96a162973b80b6ad1d96b2e1a21e6"} Dec 01 16:27:17 crc kubenswrapper[4739]: I1201 16:27:17.919310 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 01 16:27:17 crc kubenswrapper[4739]: I1201 16:27:17.993023 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 16:27:18 crc kubenswrapper[4739]: I1201 16:27:18.083693 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 01 16:27:18 crc kubenswrapper[4739]: I1201 16:27:18.146568 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 16:27:18 crc kubenswrapper[4739]: I1201 16:27:18.230834 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="3930d334-34db-4729-9064-140ba9d413f4" containerName="manila-share" containerID="cri-o://142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd" gracePeriod=30 Dec 01 16:27:18 crc kubenswrapper[4739]: I1201 16:27:18.230899 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="3930d334-34db-4729-9064-140ba9d413f4" containerName="probe" containerID="cri-o://3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298" gracePeriod=30 Dec 01 16:27:18 crc kubenswrapper[4739]: I1201 16:27:18.230981 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="5b0dcc39-d1cc-454a-a403-70d47279541c" containerName="manila-scheduler" containerID="cri-o://15897500412d62529446a37ce315bde647579396696d1eec233934952bde65a1" gracePeriod=30 Dec 01 16:27:18 crc kubenswrapper[4739]: I1201 16:27:18.231050 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="5b0dcc39-d1cc-454a-a403-70d47279541c" containerName="probe" containerID="cri-o://5df9727619fac02426e2b001a1d2dd68d53964ca1359dcd3f7b83959b9b4be85" gracePeriod=30 Dec 01 16:27:18 crc kubenswrapper[4739]: I1201 16:27:18.715102 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-8675489cb4-9mz7f" podUID="0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.240:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.240:8443: connect: connection refused" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.229097 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.243809 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7bacd19a-d9e6-4894-bc04-c11b7c5f3c85","Type":"ContainerStarted","Data":"425231dd718028d652076295d6a6b135feac3a78537583b9c2163bc3987650c0"} Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.245617 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.248982 4739 generic.go:334] "Generic (PLEG): container finished" podID="3930d334-34db-4729-9064-140ba9d413f4" containerID="3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298" exitCode=0 Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.249034 4739 generic.go:334] "Generic (PLEG): container finished" podID="3930d334-34db-4729-9064-140ba9d413f4" containerID="142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd" exitCode=1 Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.249038 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.249042 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3930d334-34db-4729-9064-140ba9d413f4","Type":"ContainerDied","Data":"3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298"} Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.249172 4739 scope.go:117] "RemoveContainer" containerID="3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.249207 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3930d334-34db-4729-9064-140ba9d413f4","Type":"ContainerDied","Data":"142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd"} Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.249236 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3930d334-34db-4729-9064-140ba9d413f4","Type":"ContainerDied","Data":"735453f371b7fdd694f79c2f8a1388f5117fb8218566102be137e523cba4a4ee"} Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.260086 4739 generic.go:334] "Generic (PLEG): container finished" podID="5b0dcc39-d1cc-454a-a403-70d47279541c" containerID="5df9727619fac02426e2b001a1d2dd68d53964ca1359dcd3f7b83959b9b4be85" exitCode=0 Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.260146 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5b0dcc39-d1cc-454a-a403-70d47279541c","Type":"ContainerDied","Data":"5df9727619fac02426e2b001a1d2dd68d53964ca1359dcd3f7b83959b9b4be85"} Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.304772 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.386046599 podStartE2EDuration="7.304744635s" podCreationTimestamp="2025-12-01 16:27:12 +0000 UTC" firstStartedPulling="2025-12-01 16:27:13.08937133 +0000 UTC m=+3134.915117444" lastFinishedPulling="2025-12-01 16:27:18.008069386 +0000 UTC m=+3139.833815480" observedRunningTime="2025-12-01 16:27:19.283687675 +0000 UTC m=+3141.109433769" watchObservedRunningTime="2025-12-01 16:27:19.304744635 +0000 UTC m=+3141.130490719" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.316706 4739 scope.go:117] "RemoveContainer" containerID="142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.345652 4739 scope.go:117] "RemoveContainer" containerID="3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298" Dec 01 16:27:19 crc kubenswrapper[4739]: E1201 16:27:19.346153 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298\": container with ID starting with 3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298 not found: ID does not exist" containerID="3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.346199 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298"} err="failed to get container status \"3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298\": rpc error: code = NotFound desc = could not find container \"3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298\": container with ID starting with 3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298 not found: ID does not exist" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.346221 4739 scope.go:117] "RemoveContainer" containerID="142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd" Dec 01 16:27:19 crc kubenswrapper[4739]: E1201 16:27:19.346592 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd\": container with ID starting with 142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd not found: ID does not exist" containerID="142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.346614 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd"} err="failed to get container status \"142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd\": rpc error: code = NotFound desc = could not find container \"142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd\": container with ID starting with 142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd not found: ID does not exist" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.346626 4739 scope.go:117] "RemoveContainer" containerID="3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.347028 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298"} err="failed to get container status \"3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298\": rpc error: code = NotFound desc = could not find container \"3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298\": container with ID starting with 3e5fd306d0acb44ef62a572098802ec1e8ebbec21c6478bf593756dc644aa298 not found: ID does not exist" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.347050 4739 scope.go:117] "RemoveContainer" containerID="142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.347278 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd"} err="failed to get container status \"142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd\": rpc error: code = NotFound desc = could not find container \"142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd\": container with ID starting with 142ec6d3e5eed008afdfe7e321ef7b20ac36a5e6bf8120b8fab2fc3e58ace7bd not found: ID does not exist" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.375073 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-combined-ca-bundle\") pod \"3930d334-34db-4729-9064-140ba9d413f4\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.375245 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-config-data-custom\") pod \"3930d334-34db-4729-9064-140ba9d413f4\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.375286 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3930d334-34db-4729-9064-140ba9d413f4-var-lib-manila\") pod \"3930d334-34db-4729-9064-140ba9d413f4\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.375349 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-scripts\") pod \"3930d334-34db-4729-9064-140ba9d413f4\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.375532 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3930d334-34db-4729-9064-140ba9d413f4-ceph\") pod \"3930d334-34db-4729-9064-140ba9d413f4\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.375563 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3930d334-34db-4729-9064-140ba9d413f4-etc-machine-id\") pod \"3930d334-34db-4729-9064-140ba9d413f4\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.375565 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3930d334-34db-4729-9064-140ba9d413f4-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "3930d334-34db-4729-9064-140ba9d413f4" (UID: "3930d334-34db-4729-9064-140ba9d413f4"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.375714 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-config-data\") pod \"3930d334-34db-4729-9064-140ba9d413f4\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.375810 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8htn\" (UniqueName: \"kubernetes.io/projected/3930d334-34db-4729-9064-140ba9d413f4-kube-api-access-q8htn\") pod \"3930d334-34db-4729-9064-140ba9d413f4\" (UID: \"3930d334-34db-4729-9064-140ba9d413f4\") " Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.375834 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3930d334-34db-4729-9064-140ba9d413f4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3930d334-34db-4729-9064-140ba9d413f4" (UID: "3930d334-34db-4729-9064-140ba9d413f4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.376827 4739 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3930d334-34db-4729-9064-140ba9d413f4-var-lib-manila\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.376851 4739 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3930d334-34db-4729-9064-140ba9d413f4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.381698 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-scripts" (OuterVolumeSpecName: "scripts") pod "3930d334-34db-4729-9064-140ba9d413f4" (UID: "3930d334-34db-4729-9064-140ba9d413f4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.382141 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3930d334-34db-4729-9064-140ba9d413f4" (UID: "3930d334-34db-4729-9064-140ba9d413f4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.382188 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3930d334-34db-4729-9064-140ba9d413f4-ceph" (OuterVolumeSpecName: "ceph") pod "3930d334-34db-4729-9064-140ba9d413f4" (UID: "3930d334-34db-4729-9064-140ba9d413f4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.382686 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3930d334-34db-4729-9064-140ba9d413f4-kube-api-access-q8htn" (OuterVolumeSpecName: "kube-api-access-q8htn") pod "3930d334-34db-4729-9064-140ba9d413f4" (UID: "3930d334-34db-4729-9064-140ba9d413f4"). InnerVolumeSpecName "kube-api-access-q8htn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.429821 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3930d334-34db-4729-9064-140ba9d413f4" (UID: "3930d334-34db-4729-9064-140ba9d413f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.477794 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:27:19 crc kubenswrapper[4739]: E1201 16:27:19.478190 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.479043 4739 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3930d334-34db-4729-9064-140ba9d413f4-ceph\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.479086 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8htn\" (UniqueName: \"kubernetes.io/projected/3930d334-34db-4729-9064-140ba9d413f4-kube-api-access-q8htn\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.479102 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.479114 4739 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.479125 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.493306 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-config-data" (OuterVolumeSpecName: "config-data") pod "3930d334-34db-4729-9064-140ba9d413f4" (UID: "3930d334-34db-4729-9064-140ba9d413f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.580852 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3930d334-34db-4729-9064-140ba9d413f4-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.592104 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.615209 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.634649 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 16:27:19 crc kubenswrapper[4739]: E1201 16:27:19.635214 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3930d334-34db-4729-9064-140ba9d413f4" containerName="probe" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.635246 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="3930d334-34db-4729-9064-140ba9d413f4" containerName="probe" Dec 01 16:27:19 crc kubenswrapper[4739]: E1201 16:27:19.635275 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3930d334-34db-4729-9064-140ba9d413f4" containerName="manila-share" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.635288 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="3930d334-34db-4729-9064-140ba9d413f4" containerName="manila-share" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.635585 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="3930d334-34db-4729-9064-140ba9d413f4" containerName="probe" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.635633 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="3930d334-34db-4729-9064-140ba9d413f4" containerName="manila-share" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.637269 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.657933 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.660024 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.785499 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf03c1ab-74fb-412e-9625-3993db3e51df-scripts\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.785546 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf03c1ab-74fb-412e-9625-3993db3e51df-config-data\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.785601 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wshrc\" (UniqueName: \"kubernetes.io/projected/cf03c1ab-74fb-412e-9625-3993db3e51df-kube-api-access-wshrc\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.785625 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf03c1ab-74fb-412e-9625-3993db3e51df-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.786095 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cf03c1ab-74fb-412e-9625-3993db3e51df-ceph\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.786378 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cf03c1ab-74fb-412e-9625-3993db3e51df-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.786489 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/cf03c1ab-74fb-412e-9625-3993db3e51df-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.786565 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cf03c1ab-74fb-412e-9625-3993db3e51df-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.889629 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf03c1ab-74fb-412e-9625-3993db3e51df-scripts\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.889714 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf03c1ab-74fb-412e-9625-3993db3e51df-config-data\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.889810 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wshrc\" (UniqueName: \"kubernetes.io/projected/cf03c1ab-74fb-412e-9625-3993db3e51df-kube-api-access-wshrc\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.889846 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf03c1ab-74fb-412e-9625-3993db3e51df-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.890555 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cf03c1ab-74fb-412e-9625-3993db3e51df-ceph\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.890658 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cf03c1ab-74fb-412e-9625-3993db3e51df-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.890750 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/cf03c1ab-74fb-412e-9625-3993db3e51df-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.890832 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cf03c1ab-74fb-412e-9625-3993db3e51df-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.890828 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cf03c1ab-74fb-412e-9625-3993db3e51df-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.890980 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/cf03c1ab-74fb-412e-9625-3993db3e51df-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.893475 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf03c1ab-74fb-412e-9625-3993db3e51df-scripts\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.893590 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf03c1ab-74fb-412e-9625-3993db3e51df-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.895153 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf03c1ab-74fb-412e-9625-3993db3e51df-config-data\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.895879 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cf03c1ab-74fb-412e-9625-3993db3e51df-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.905703 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cf03c1ab-74fb-412e-9625-3993db3e51df-ceph\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.926785 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wshrc\" (UniqueName: \"kubernetes.io/projected/cf03c1ab-74fb-412e-9625-3993db3e51df-kube-api-access-wshrc\") pod \"manila-share-share1-0\" (UID: \"cf03c1ab-74fb-412e-9625-3993db3e51df\") " pod="openstack/manila-share-share1-0" Dec 01 16:27:19 crc kubenswrapper[4739]: I1201 16:27:19.966363 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 01 16:27:20 crc kubenswrapper[4739]: I1201 16:27:20.488128 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3930d334-34db-4729-9064-140ba9d413f4" path="/var/lib/kubelet/pods/3930d334-34db-4729-9064-140ba9d413f4/volumes" Dec 01 16:27:20 crc kubenswrapper[4739]: I1201 16:27:20.558815 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 01 16:27:20 crc kubenswrapper[4739]: W1201 16:27:20.567151 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf03c1ab_74fb_412e_9625_3993db3e51df.slice/crio-af3ea33771990ae226c4191cc0c057011f6b500627f1639fcfd822db5dd990a8 WatchSource:0}: Error finding container af3ea33771990ae226c4191cc0c057011f6b500627f1639fcfd822db5dd990a8: Status 404 returned error can't find the container with id af3ea33771990ae226c4191cc0c057011f6b500627f1639fcfd822db5dd990a8 Dec 01 16:27:21 crc kubenswrapper[4739]: I1201 16:27:21.296395 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"cf03c1ab-74fb-412e-9625-3993db3e51df","Type":"ContainerStarted","Data":"1670cb6a6788debe385fad1da4428cd2ede7196818de12c91b506e9c4869416f"} Dec 01 16:27:21 crc kubenswrapper[4739]: I1201 16:27:21.296696 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"cf03c1ab-74fb-412e-9625-3993db3e51df","Type":"ContainerStarted","Data":"af3ea33771990ae226c4191cc0c057011f6b500627f1639fcfd822db5dd990a8"} Dec 01 16:27:22 crc kubenswrapper[4739]: I1201 16:27:22.305293 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"cf03c1ab-74fb-412e-9625-3993db3e51df","Type":"ContainerStarted","Data":"3c995b2da7ec33dc146135fc185aca666b0ebfdd803ab9e0d208f7fb1a050d55"} Dec 01 16:27:22 crc kubenswrapper[4739]: I1201 16:27:22.332713 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.3326856449999998 podStartE2EDuration="3.332685645s" podCreationTimestamp="2025-12-01 16:27:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:27:22.320401475 +0000 UTC m=+3144.146147589" watchObservedRunningTime="2025-12-01 16:27:22.332685645 +0000 UTC m=+3144.158431759" Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.315375 4739 generic.go:334] "Generic (PLEG): container finished" podID="5b0dcc39-d1cc-454a-a403-70d47279541c" containerID="15897500412d62529446a37ce315bde647579396696d1eec233934952bde65a1" exitCode=0 Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.315463 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5b0dcc39-d1cc-454a-a403-70d47279541c","Type":"ContainerDied","Data":"15897500412d62529446a37ce315bde647579396696d1eec233934952bde65a1"} Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.315774 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5b0dcc39-d1cc-454a-a403-70d47279541c","Type":"ContainerDied","Data":"a571b6c958638eafcef691e477afcd94f108497774c3aff88d1c6a7bba8d7d4b"} Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.315791 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a571b6c958638eafcef691e477afcd94f108497774c3aff88d1c6a7bba8d7d4b" Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.337457 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.409320 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b0dcc39-d1cc-454a-a403-70d47279541c-etc-machine-id\") pod \"5b0dcc39-d1cc-454a-a403-70d47279541c\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.409482 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-config-data\") pod \"5b0dcc39-d1cc-454a-a403-70d47279541c\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.409526 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dhqd\" (UniqueName: \"kubernetes.io/projected/5b0dcc39-d1cc-454a-a403-70d47279541c-kube-api-access-7dhqd\") pod \"5b0dcc39-d1cc-454a-a403-70d47279541c\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.409697 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b0dcc39-d1cc-454a-a403-70d47279541c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5b0dcc39-d1cc-454a-a403-70d47279541c" (UID: "5b0dcc39-d1cc-454a-a403-70d47279541c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.409716 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-config-data-custom\") pod \"5b0dcc39-d1cc-454a-a403-70d47279541c\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.409966 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-combined-ca-bundle\") pod \"5b0dcc39-d1cc-454a-a403-70d47279541c\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.410058 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-scripts\") pod \"5b0dcc39-d1cc-454a-a403-70d47279541c\" (UID: \"5b0dcc39-d1cc-454a-a403-70d47279541c\") " Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.410790 4739 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b0dcc39-d1cc-454a-a403-70d47279541c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.416100 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5b0dcc39-d1cc-454a-a403-70d47279541c" (UID: "5b0dcc39-d1cc-454a-a403-70d47279541c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.416159 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b0dcc39-d1cc-454a-a403-70d47279541c-kube-api-access-7dhqd" (OuterVolumeSpecName: "kube-api-access-7dhqd") pod "5b0dcc39-d1cc-454a-a403-70d47279541c" (UID: "5b0dcc39-d1cc-454a-a403-70d47279541c"). InnerVolumeSpecName "kube-api-access-7dhqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.418805 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-scripts" (OuterVolumeSpecName: "scripts") pod "5b0dcc39-d1cc-454a-a403-70d47279541c" (UID: "5b0dcc39-d1cc-454a-a403-70d47279541c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.471257 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b0dcc39-d1cc-454a-a403-70d47279541c" (UID: "5b0dcc39-d1cc-454a-a403-70d47279541c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.512769 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dhqd\" (UniqueName: \"kubernetes.io/projected/5b0dcc39-d1cc-454a-a403-70d47279541c-kube-api-access-7dhqd\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.512805 4739 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.512818 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.512834 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.517010 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-config-data" (OuterVolumeSpecName: "config-data") pod "5b0dcc39-d1cc-454a-a403-70d47279541c" (UID: "5b0dcc39-d1cc-454a-a403-70d47279541c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:23 crc kubenswrapper[4739]: I1201 16:27:23.613568 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b0dcc39-d1cc-454a-a403-70d47279541c-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.127873 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.332624 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.377516 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.387615 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.399518 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 16:27:24 crc kubenswrapper[4739]: E1201 16:27:24.400118 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0dcc39-d1cc-454a-a403-70d47279541c" containerName="manila-scheduler" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.400145 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0dcc39-d1cc-454a-a403-70d47279541c" containerName="manila-scheduler" Dec 01 16:27:24 crc kubenswrapper[4739]: E1201 16:27:24.400207 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0dcc39-d1cc-454a-a403-70d47279541c" containerName="probe" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.400219 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0dcc39-d1cc-454a-a403-70d47279541c" containerName="probe" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.400532 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0dcc39-d1cc-454a-a403-70d47279541c" containerName="probe" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.400643 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0dcc39-d1cc-454a-a403-70d47279541c" containerName="manila-scheduler" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.406286 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.409645 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.412789 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.432106 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be94c9f8-1f60-4cd9-b6f7-25459823386c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.432242 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh4rj\" (UniqueName: \"kubernetes.io/projected/be94c9f8-1f60-4cd9-b6f7-25459823386c-kube-api-access-kh4rj\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.432320 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be94c9f8-1f60-4cd9-b6f7-25459823386c-config-data\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.432473 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be94c9f8-1f60-4cd9-b6f7-25459823386c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.432623 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be94c9f8-1f60-4cd9-b6f7-25459823386c-scripts\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.432721 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be94c9f8-1f60-4cd9-b6f7-25459823386c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.490580 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b0dcc39-d1cc-454a-a403-70d47279541c" path="/var/lib/kubelet/pods/5b0dcc39-d1cc-454a-a403-70d47279541c/volumes" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.534320 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be94c9f8-1f60-4cd9-b6f7-25459823386c-scripts\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.534530 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be94c9f8-1f60-4cd9-b6f7-25459823386c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.535462 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be94c9f8-1f60-4cd9-b6f7-25459823386c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.535621 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh4rj\" (UniqueName: \"kubernetes.io/projected/be94c9f8-1f60-4cd9-b6f7-25459823386c-kube-api-access-kh4rj\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.535765 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be94c9f8-1f60-4cd9-b6f7-25459823386c-config-data\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.535941 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be94c9f8-1f60-4cd9-b6f7-25459823386c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.536126 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be94c9f8-1f60-4cd9-b6f7-25459823386c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.540154 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be94c9f8-1f60-4cd9-b6f7-25459823386c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.541206 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be94c9f8-1f60-4cd9-b6f7-25459823386c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.541645 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be94c9f8-1f60-4cd9-b6f7-25459823386c-scripts\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.543681 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be94c9f8-1f60-4cd9-b6f7-25459823386c-config-data\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.559947 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh4rj\" (UniqueName: \"kubernetes.io/projected/be94c9f8-1f60-4cd9-b6f7-25459823386c-kube-api-access-kh4rj\") pod \"manila-scheduler-0\" (UID: \"be94c9f8-1f60-4cd9-b6f7-25459823386c\") " pod="openstack/manila-scheduler-0" Dec 01 16:27:24 crc kubenswrapper[4739]: I1201 16:27:24.730369 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 01 16:27:25 crc kubenswrapper[4739]: I1201 16:27:25.238511 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 01 16:27:25 crc kubenswrapper[4739]: I1201 16:27:25.354562 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"be94c9f8-1f60-4cd9-b6f7-25459823386c","Type":"ContainerStarted","Data":"d88e77f349cf87db9c91a6647c80ed668f01cba7f97e330801843962bfe47580"} Dec 01 16:27:26 crc kubenswrapper[4739]: I1201 16:27:26.369893 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"be94c9f8-1f60-4cd9-b6f7-25459823386c","Type":"ContainerStarted","Data":"4885c734caf079feffc20524a02b5f83b05e7180b20c76238b1ae6942910ee76"} Dec 01 16:27:26 crc kubenswrapper[4739]: I1201 16:27:26.370202 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"be94c9f8-1f60-4cd9-b6f7-25459823386c","Type":"ContainerStarted","Data":"b685cffd9ec68136fff32e97ca65a255c252a649cbafa5934c4e55b63f2390bb"} Dec 01 16:27:26 crc kubenswrapper[4739]: I1201 16:27:26.399075 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.399058353 podStartE2EDuration="2.399058353s" podCreationTimestamp="2025-12-01 16:27:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:27:26.394720909 +0000 UTC m=+3148.220467013" watchObservedRunningTime="2025-12-01 16:27:26.399058353 +0000 UTC m=+3148.224804467" Dec 01 16:27:28 crc kubenswrapper[4739]: I1201 16:27:28.715665 4739 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-8675489cb4-9mz7f" podUID="0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.240:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.240:8443: connect: connection refused" Dec 01 16:27:28 crc kubenswrapper[4739]: I1201 16:27:28.718266 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:27:29 crc kubenswrapper[4739]: I1201 16:27:29.967506 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 01 16:27:30 crc kubenswrapper[4739]: I1201 16:27:30.477977 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:27:30 crc kubenswrapper[4739]: E1201 16:27:30.478795 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:27:33 crc kubenswrapper[4739]: I1201 16:27:33.895008 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hf8ns"] Dec 01 16:27:33 crc kubenswrapper[4739]: I1201 16:27:33.900306 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hf8ns" Dec 01 16:27:33 crc kubenswrapper[4739]: I1201 16:27:33.915324 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hf8ns"] Dec 01 16:27:33 crc kubenswrapper[4739]: I1201 16:27:33.993051 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpwwl\" (UniqueName: \"kubernetes.io/projected/cdcee334-c748-423c-ae7b-a4de1795a21b-kube-api-access-qpwwl\") pod \"community-operators-hf8ns\" (UID: \"cdcee334-c748-423c-ae7b-a4de1795a21b\") " pod="openshift-marketplace/community-operators-hf8ns" Dec 01 16:27:33 crc kubenswrapper[4739]: I1201 16:27:33.993100 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdcee334-c748-423c-ae7b-a4de1795a21b-utilities\") pod \"community-operators-hf8ns\" (UID: \"cdcee334-c748-423c-ae7b-a4de1795a21b\") " pod="openshift-marketplace/community-operators-hf8ns" Dec 01 16:27:33 crc kubenswrapper[4739]: I1201 16:27:33.993330 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdcee334-c748-423c-ae7b-a4de1795a21b-catalog-content\") pod \"community-operators-hf8ns\" (UID: \"cdcee334-c748-423c-ae7b-a4de1795a21b\") " pod="openshift-marketplace/community-operators-hf8ns" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.097167 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpwwl\" (UniqueName: \"kubernetes.io/projected/cdcee334-c748-423c-ae7b-a4de1795a21b-kube-api-access-qpwwl\") pod \"community-operators-hf8ns\" (UID: \"cdcee334-c748-423c-ae7b-a4de1795a21b\") " pod="openshift-marketplace/community-operators-hf8ns" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.097247 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdcee334-c748-423c-ae7b-a4de1795a21b-utilities\") pod \"community-operators-hf8ns\" (UID: \"cdcee334-c748-423c-ae7b-a4de1795a21b\") " pod="openshift-marketplace/community-operators-hf8ns" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.097352 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdcee334-c748-423c-ae7b-a4de1795a21b-catalog-content\") pod \"community-operators-hf8ns\" (UID: \"cdcee334-c748-423c-ae7b-a4de1795a21b\") " pod="openshift-marketplace/community-operators-hf8ns" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.098469 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdcee334-c748-423c-ae7b-a4de1795a21b-catalog-content\") pod \"community-operators-hf8ns\" (UID: \"cdcee334-c748-423c-ae7b-a4de1795a21b\") " pod="openshift-marketplace/community-operators-hf8ns" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.098945 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdcee334-c748-423c-ae7b-a4de1795a21b-utilities\") pod \"community-operators-hf8ns\" (UID: \"cdcee334-c748-423c-ae7b-a4de1795a21b\") " pod="openshift-marketplace/community-operators-hf8ns" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.127604 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpwwl\" (UniqueName: \"kubernetes.io/projected/cdcee334-c748-423c-ae7b-a4de1795a21b-kube-api-access-qpwwl\") pod \"community-operators-hf8ns\" (UID: \"cdcee334-c748-423c-ae7b-a4de1795a21b\") " pod="openshift-marketplace/community-operators-hf8ns" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.236838 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hf8ns" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.468586 4739 generic.go:334] "Generic (PLEG): container finished" podID="0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" containerID="82e1bf034e64723a1b9fead779cd64592427fad19be93d92b9185d673b6e38ed" exitCode=137 Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.468618 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8675489cb4-9mz7f" event={"ID":"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7","Type":"ContainerDied","Data":"82e1bf034e64723a1b9fead779cd64592427fad19be93d92b9185d673b6e38ed"} Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.615277 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.712058 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-combined-ca-bundle\") pod \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.715736 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-config-data\") pod \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.715780 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmg78\" (UniqueName: \"kubernetes.io/projected/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-kube-api-access-kmg78\") pod \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.715877 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-logs\") pod \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.715957 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-horizon-secret-key\") pod \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.715994 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-horizon-tls-certs\") pod \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.716041 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-scripts\") pod \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\" (UID: \"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7\") " Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.716513 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-logs" (OuterVolumeSpecName: "logs") pod "0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" (UID: "0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.716961 4739 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.720236 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-kube-api-access-kmg78" (OuterVolumeSpecName: "kube-api-access-kmg78") pod "0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" (UID: "0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7"). InnerVolumeSpecName "kube-api-access-kmg78". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.722510 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" (UID: "0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.730975 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.740691 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-scripts" (OuterVolumeSpecName: "scripts") pod "0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" (UID: "0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.740711 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-config-data" (OuterVolumeSpecName: "config-data") pod "0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" (UID: "0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.749657 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" (UID: "0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.766307 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" (UID: "0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.819341 4739 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.819384 4739 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.819397 4739 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.819406 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.819437 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.819446 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmg78\" (UniqueName: \"kubernetes.io/projected/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7-kube-api-access-kmg78\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:34 crc kubenswrapper[4739]: I1201 16:27:34.881672 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hf8ns"] Dec 01 16:27:35 crc kubenswrapper[4739]: I1201 16:27:35.489460 4739 generic.go:334] "Generic (PLEG): container finished" podID="cdcee334-c748-423c-ae7b-a4de1795a21b" containerID="cd2695c65e38edb48b451245b9ff7b9ae5e4611247077651ff8a3cb2bced2723" exitCode=0 Dec 01 16:27:35 crc kubenswrapper[4739]: I1201 16:27:35.489590 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf8ns" event={"ID":"cdcee334-c748-423c-ae7b-a4de1795a21b","Type":"ContainerDied","Data":"cd2695c65e38edb48b451245b9ff7b9ae5e4611247077651ff8a3cb2bced2723"} Dec 01 16:27:35 crc kubenswrapper[4739]: I1201 16:27:35.490573 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf8ns" event={"ID":"cdcee334-c748-423c-ae7b-a4de1795a21b","Type":"ContainerStarted","Data":"0f69994844afb882e22907823885ec82fb6137f70002dd4ee48ff68800bbc08c"} Dec 01 16:27:35 crc kubenswrapper[4739]: I1201 16:27:35.497726 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8675489cb4-9mz7f" event={"ID":"0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7","Type":"ContainerDied","Data":"250d778ece47f00a405232d3d1cd1bd59cd0ad8ebd1397ae09658d4a7bdd1a2e"} Dec 01 16:27:35 crc kubenswrapper[4739]: I1201 16:27:35.497783 4739 scope.go:117] "RemoveContainer" containerID="c30f8e2e727a49a2094463f0e8cf3a19335dd473e4173694417d8ca96e055330" Dec 01 16:27:35 crc kubenswrapper[4739]: I1201 16:27:35.497971 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8675489cb4-9mz7f" Dec 01 16:27:35 crc kubenswrapper[4739]: I1201 16:27:35.552691 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8675489cb4-9mz7f"] Dec 01 16:27:35 crc kubenswrapper[4739]: I1201 16:27:35.560846 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-8675489cb4-9mz7f"] Dec 01 16:27:35 crc kubenswrapper[4739]: I1201 16:27:35.688803 4739 scope.go:117] "RemoveContainer" containerID="82e1bf034e64723a1b9fead779cd64592427fad19be93d92b9185d673b6e38ed" Dec 01 16:27:36 crc kubenswrapper[4739]: I1201 16:27:36.503378 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" path="/var/lib/kubelet/pods/0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7/volumes" Dec 01 16:27:40 crc kubenswrapper[4739]: I1201 16:27:40.563537 4739 generic.go:334] "Generic (PLEG): container finished" podID="cdcee334-c748-423c-ae7b-a4de1795a21b" containerID="d9e275281cae51d28d87581fa5580b49b242cfe615fe6d4c53cc497dd9bd41a3" exitCode=0 Dec 01 16:27:40 crc kubenswrapper[4739]: I1201 16:27:40.563603 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf8ns" event={"ID":"cdcee334-c748-423c-ae7b-a4de1795a21b","Type":"ContainerDied","Data":"d9e275281cae51d28d87581fa5580b49b242cfe615fe6d4c53cc497dd9bd41a3"} Dec 01 16:27:42 crc kubenswrapper[4739]: I1201 16:27:42.321064 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 01 16:27:42 crc kubenswrapper[4739]: I1201 16:27:42.551016 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 16:27:42 crc kubenswrapper[4739]: I1201 16:27:42.582946 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf8ns" event={"ID":"cdcee334-c748-423c-ae7b-a4de1795a21b","Type":"ContainerStarted","Data":"6c34059b7fad65db769b24f925cddc0787eaf7f7fd59bed495bda678cd5ac556"} Dec 01 16:27:42 crc kubenswrapper[4739]: I1201 16:27:42.611018 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hf8ns" podStartSLOduration=3.259588093 podStartE2EDuration="9.610996593s" podCreationTimestamp="2025-12-01 16:27:33 +0000 UTC" firstStartedPulling="2025-12-01 16:27:35.490850657 +0000 UTC m=+3157.316596741" lastFinishedPulling="2025-12-01 16:27:41.842259097 +0000 UTC m=+3163.668005241" observedRunningTime="2025-12-01 16:27:42.609178437 +0000 UTC m=+3164.434924551" watchObservedRunningTime="2025-12-01 16:27:42.610996593 +0000 UTC m=+3164.436742687" Dec 01 16:27:43 crc kubenswrapper[4739]: I1201 16:27:43.477564 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:27:44 crc kubenswrapper[4739]: I1201 16:27:44.238331 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hf8ns" Dec 01 16:27:44 crc kubenswrapper[4739]: I1201 16:27:44.238708 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hf8ns" Dec 01 16:27:44 crc kubenswrapper[4739]: I1201 16:27:44.605824 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"8c69d57ff87a87518f586ab31047e56d7736ae4a7c0a6aae691054d7b0c41c0d"} Dec 01 16:27:45 crc kubenswrapper[4739]: I1201 16:27:45.287671 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-hf8ns" podUID="cdcee334-c748-423c-ae7b-a4de1795a21b" containerName="registry-server" probeResult="failure" output=< Dec 01 16:27:45 crc kubenswrapper[4739]: timeout: failed to connect service ":50051" within 1s Dec 01 16:27:45 crc kubenswrapper[4739]: > Dec 01 16:27:46 crc kubenswrapper[4739]: I1201 16:27:46.321360 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 01 16:27:54 crc kubenswrapper[4739]: I1201 16:27:54.302363 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hf8ns" Dec 01 16:27:54 crc kubenswrapper[4739]: I1201 16:27:54.387312 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hf8ns" Dec 01 16:27:54 crc kubenswrapper[4739]: I1201 16:27:54.464449 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hf8ns"] Dec 01 16:27:54 crc kubenswrapper[4739]: I1201 16:27:54.557322 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-52jpf"] Dec 01 16:27:54 crc kubenswrapper[4739]: I1201 16:27:54.557728 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-52jpf" podUID="079aa9ef-3896-4f11-9a9f-e8cb26607b88" containerName="registry-server" containerID="cri-o://77db31bd9f4b2ea11b8cb617f98352ce305dbc6cf270e1306446874327e7d00d" gracePeriod=2 Dec 01 16:27:54 crc kubenswrapper[4739]: I1201 16:27:54.724598 4739 generic.go:334] "Generic (PLEG): container finished" podID="079aa9ef-3896-4f11-9a9f-e8cb26607b88" containerID="77db31bd9f4b2ea11b8cb617f98352ce305dbc6cf270e1306446874327e7d00d" exitCode=0 Dec 01 16:27:54 crc kubenswrapper[4739]: I1201 16:27:54.724661 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52jpf" event={"ID":"079aa9ef-3896-4f11-9a9f-e8cb26607b88","Type":"ContainerDied","Data":"77db31bd9f4b2ea11b8cb617f98352ce305dbc6cf270e1306446874327e7d00d"} Dec 01 16:27:54 crc kubenswrapper[4739]: I1201 16:27:54.995111 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-52jpf" Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.055980 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7tvw\" (UniqueName: \"kubernetes.io/projected/079aa9ef-3896-4f11-9a9f-e8cb26607b88-kube-api-access-v7tvw\") pod \"079aa9ef-3896-4f11-9a9f-e8cb26607b88\" (UID: \"079aa9ef-3896-4f11-9a9f-e8cb26607b88\") " Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.056034 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/079aa9ef-3896-4f11-9a9f-e8cb26607b88-utilities\") pod \"079aa9ef-3896-4f11-9a9f-e8cb26607b88\" (UID: \"079aa9ef-3896-4f11-9a9f-e8cb26607b88\") " Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.056212 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/079aa9ef-3896-4f11-9a9f-e8cb26607b88-catalog-content\") pod \"079aa9ef-3896-4f11-9a9f-e8cb26607b88\" (UID: \"079aa9ef-3896-4f11-9a9f-e8cb26607b88\") " Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.056753 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/079aa9ef-3896-4f11-9a9f-e8cb26607b88-utilities" (OuterVolumeSpecName: "utilities") pod "079aa9ef-3896-4f11-9a9f-e8cb26607b88" (UID: "079aa9ef-3896-4f11-9a9f-e8cb26607b88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.062779 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/079aa9ef-3896-4f11-9a9f-e8cb26607b88-kube-api-access-v7tvw" (OuterVolumeSpecName: "kube-api-access-v7tvw") pod "079aa9ef-3896-4f11-9a9f-e8cb26607b88" (UID: "079aa9ef-3896-4f11-9a9f-e8cb26607b88"). InnerVolumeSpecName "kube-api-access-v7tvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.124128 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/079aa9ef-3896-4f11-9a9f-e8cb26607b88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "079aa9ef-3896-4f11-9a9f-e8cb26607b88" (UID: "079aa9ef-3896-4f11-9a9f-e8cb26607b88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.159122 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7tvw\" (UniqueName: \"kubernetes.io/projected/079aa9ef-3896-4f11-9a9f-e8cb26607b88-kube-api-access-v7tvw\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.159158 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/079aa9ef-3896-4f11-9a9f-e8cb26607b88-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.159171 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/079aa9ef-3896-4f11-9a9f-e8cb26607b88-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.741770 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-52jpf" Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.741969 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-52jpf" event={"ID":"079aa9ef-3896-4f11-9a9f-e8cb26607b88","Type":"ContainerDied","Data":"c7d89dc3eb3eaa062e4982aa42686cc4051cd923af4fccac4119629a6caff56d"} Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.742232 4739 scope.go:117] "RemoveContainer" containerID="77db31bd9f4b2ea11b8cb617f98352ce305dbc6cf270e1306446874327e7d00d" Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.793398 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-52jpf"] Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.797143 4739 scope.go:117] "RemoveContainer" containerID="efd49d28edb13c19a93d0059d68d8e724e4724e47798fe264251652b4edd22f6" Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.802814 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-52jpf"] Dec 01 16:27:55 crc kubenswrapper[4739]: I1201 16:27:55.823629 4739 scope.go:117] "RemoveContainer" containerID="4391f9750ac51234dc0a16cfb7b8e2c1938e004783714af7ec9bc68170f87edd" Dec 01 16:27:56 crc kubenswrapper[4739]: I1201 16:27:56.499040 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="079aa9ef-3896-4f11-9a9f-e8cb26607b88" path="/var/lib/kubelet/pods/079aa9ef-3896-4f11-9a9f-e8cb26607b88/volumes" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.438148 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 01 16:28:42 crc kubenswrapper[4739]: E1201 16:28:42.440495 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="079aa9ef-3896-4f11-9a9f-e8cb26607b88" containerName="registry-server" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.440537 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="079aa9ef-3896-4f11-9a9f-e8cb26607b88" containerName="registry-server" Dec 01 16:28:42 crc kubenswrapper[4739]: E1201 16:28:42.440567 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" containerName="horizon" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.440580 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" containerName="horizon" Dec 01 16:28:42 crc kubenswrapper[4739]: E1201 16:28:42.440614 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" containerName="horizon-log" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.440629 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" containerName="horizon-log" Dec 01 16:28:42 crc kubenswrapper[4739]: E1201 16:28:42.440645 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="079aa9ef-3896-4f11-9a9f-e8cb26607b88" containerName="extract-content" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.440658 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="079aa9ef-3896-4f11-9a9f-e8cb26607b88" containerName="extract-content" Dec 01 16:28:42 crc kubenswrapper[4739]: E1201 16:28:42.440676 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="079aa9ef-3896-4f11-9a9f-e8cb26607b88" containerName="extract-utilities" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.440687 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="079aa9ef-3896-4f11-9a9f-e8cb26607b88" containerName="extract-utilities" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.441011 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" containerName="horizon" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.441051 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="079aa9ef-3896-4f11-9a9f-e8cb26607b88" containerName="registry-server" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.441070 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dd0df7f-ae84-4ae8-8e0d-bdf5941b4ce7" containerName="horizon-log" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.442351 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.447990 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.448099 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.448238 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.449217 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cpqn6" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.458832 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.568619 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.568690 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.568764 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvzjb\" (UniqueName: \"kubernetes.io/projected/d7438aa0-1e45-4112-9af4-a584825d29df-kube-api-access-cvzjb\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.568848 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.568878 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.569171 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7438aa0-1e45-4112-9af4-a584825d29df-config-data\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.569275 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d7438aa0-1e45-4112-9af4-a584825d29df-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.569343 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d7438aa0-1e45-4112-9af4-a584825d29df-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.569714 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d7438aa0-1e45-4112-9af4-a584825d29df-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.672087 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7438aa0-1e45-4112-9af4-a584825d29df-config-data\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.672195 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d7438aa0-1e45-4112-9af4-a584825d29df-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.672251 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d7438aa0-1e45-4112-9af4-a584825d29df-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.672388 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d7438aa0-1e45-4112-9af4-a584825d29df-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.672508 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.672551 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.672613 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvzjb\" (UniqueName: \"kubernetes.io/projected/d7438aa0-1e45-4112-9af4-a584825d29df-kube-api-access-cvzjb\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.672691 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.672731 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.673172 4739 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.673815 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d7438aa0-1e45-4112-9af4-a584825d29df-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.674245 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d7438aa0-1e45-4112-9af4-a584825d29df-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.674315 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d7438aa0-1e45-4112-9af4-a584825d29df-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.675044 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7438aa0-1e45-4112-9af4-a584825d29df-config-data\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.683572 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.684073 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.684402 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.697548 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvzjb\" (UniqueName: \"kubernetes.io/projected/d7438aa0-1e45-4112-9af4-a584825d29df-kube-api-access-cvzjb\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.733089 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " pod="openstack/tempest-tests-tempest" Dec 01 16:28:42 crc kubenswrapper[4739]: I1201 16:28:42.787189 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 16:28:43 crc kubenswrapper[4739]: I1201 16:28:43.314228 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 01 16:28:44 crc kubenswrapper[4739]: I1201 16:28:44.281903 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d7438aa0-1e45-4112-9af4-a584825d29df","Type":"ContainerStarted","Data":"2492283905a5c55c900b3eba6a0122b3f3c9e6da2f0de8d0cfa8fcd7d80610e9"} Dec 01 16:29:11 crc kubenswrapper[4739]: I1201 16:29:11.060197 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8zzt6"] Dec 01 16:29:11 crc kubenswrapper[4739]: I1201 16:29:11.063266 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:11 crc kubenswrapper[4739]: I1201 16:29:11.098411 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8zzt6"] Dec 01 16:29:11 crc kubenswrapper[4739]: I1201 16:29:11.151931 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-catalog-content\") pod \"redhat-operators-8zzt6\" (UID: \"fa52dc96-cd9d-4e00-b777-6d345d53e3d6\") " pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:11 crc kubenswrapper[4739]: I1201 16:29:11.151997 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlh7r\" (UniqueName: \"kubernetes.io/projected/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-kube-api-access-xlh7r\") pod \"redhat-operators-8zzt6\" (UID: \"fa52dc96-cd9d-4e00-b777-6d345d53e3d6\") " pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:11 crc kubenswrapper[4739]: I1201 16:29:11.152040 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-utilities\") pod \"redhat-operators-8zzt6\" (UID: \"fa52dc96-cd9d-4e00-b777-6d345d53e3d6\") " pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:11 crc kubenswrapper[4739]: I1201 16:29:11.254499 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-catalog-content\") pod \"redhat-operators-8zzt6\" (UID: \"fa52dc96-cd9d-4e00-b777-6d345d53e3d6\") " pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:11 crc kubenswrapper[4739]: I1201 16:29:11.254575 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlh7r\" (UniqueName: \"kubernetes.io/projected/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-kube-api-access-xlh7r\") pod \"redhat-operators-8zzt6\" (UID: \"fa52dc96-cd9d-4e00-b777-6d345d53e3d6\") " pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:11 crc kubenswrapper[4739]: I1201 16:29:11.254600 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-utilities\") pod \"redhat-operators-8zzt6\" (UID: \"fa52dc96-cd9d-4e00-b777-6d345d53e3d6\") " pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:11 crc kubenswrapper[4739]: I1201 16:29:11.255087 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-catalog-content\") pod \"redhat-operators-8zzt6\" (UID: \"fa52dc96-cd9d-4e00-b777-6d345d53e3d6\") " pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:11 crc kubenswrapper[4739]: I1201 16:29:11.255092 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-utilities\") pod \"redhat-operators-8zzt6\" (UID: \"fa52dc96-cd9d-4e00-b777-6d345d53e3d6\") " pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:11 crc kubenswrapper[4739]: I1201 16:29:11.274250 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlh7r\" (UniqueName: \"kubernetes.io/projected/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-kube-api-access-xlh7r\") pod \"redhat-operators-8zzt6\" (UID: \"fa52dc96-cd9d-4e00-b777-6d345d53e3d6\") " pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:11 crc kubenswrapper[4739]: I1201 16:29:11.409223 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:18 crc kubenswrapper[4739]: E1201 16:29:18.485244 4739 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 01 16:29:18 crc kubenswrapper[4739]: E1201 16:29:18.485866 4739 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cvzjb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(d7438aa0-1e45-4112-9af4-a584825d29df): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 16:29:18 crc kubenswrapper[4739]: E1201 16:29:18.487032 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="d7438aa0-1e45-4112-9af4-a584825d29df" Dec 01 16:29:18 crc kubenswrapper[4739]: E1201 16:29:18.642999 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="d7438aa0-1e45-4112-9af4-a584825d29df" Dec 01 16:29:18 crc kubenswrapper[4739]: I1201 16:29:18.931379 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8zzt6"] Dec 01 16:29:18 crc kubenswrapper[4739]: W1201 16:29:18.938852 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa52dc96_cd9d_4e00_b777_6d345d53e3d6.slice/crio-c28e5d01681df4d36ea18d0cd759c8934039f83ea7a515ee3d72fc539b430f96 WatchSource:0}: Error finding container c28e5d01681df4d36ea18d0cd759c8934039f83ea7a515ee3d72fc539b430f96: Status 404 returned error can't find the container with id c28e5d01681df4d36ea18d0cd759c8934039f83ea7a515ee3d72fc539b430f96 Dec 01 16:29:19 crc kubenswrapper[4739]: I1201 16:29:19.653813 4739 generic.go:334] "Generic (PLEG): container finished" podID="fa52dc96-cd9d-4e00-b777-6d345d53e3d6" containerID="3b5c0da23349c43b19a5fcd17df12430faa26f597de7c54a842a4cfafe4c9b45" exitCode=0 Dec 01 16:29:19 crc kubenswrapper[4739]: I1201 16:29:19.654017 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zzt6" event={"ID":"fa52dc96-cd9d-4e00-b777-6d345d53e3d6","Type":"ContainerDied","Data":"3b5c0da23349c43b19a5fcd17df12430faa26f597de7c54a842a4cfafe4c9b45"} Dec 01 16:29:19 crc kubenswrapper[4739]: I1201 16:29:19.654244 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zzt6" event={"ID":"fa52dc96-cd9d-4e00-b777-6d345d53e3d6","Type":"ContainerStarted","Data":"c28e5d01681df4d36ea18d0cd759c8934039f83ea7a515ee3d72fc539b430f96"} Dec 01 16:29:20 crc kubenswrapper[4739]: I1201 16:29:20.665757 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zzt6" event={"ID":"fa52dc96-cd9d-4e00-b777-6d345d53e3d6","Type":"ContainerStarted","Data":"0b4f60433ea97f6a6645239b4b128d33d0d368d5a3da6e765b8261bf5d61f280"} Dec 01 16:29:22 crc kubenswrapper[4739]: I1201 16:29:22.682951 4739 generic.go:334] "Generic (PLEG): container finished" podID="fa52dc96-cd9d-4e00-b777-6d345d53e3d6" containerID="0b4f60433ea97f6a6645239b4b128d33d0d368d5a3da6e765b8261bf5d61f280" exitCode=0 Dec 01 16:29:22 crc kubenswrapper[4739]: I1201 16:29:22.683030 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zzt6" event={"ID":"fa52dc96-cd9d-4e00-b777-6d345d53e3d6","Type":"ContainerDied","Data":"0b4f60433ea97f6a6645239b4b128d33d0d368d5a3da6e765b8261bf5d61f280"} Dec 01 16:29:23 crc kubenswrapper[4739]: I1201 16:29:23.694745 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zzt6" event={"ID":"fa52dc96-cd9d-4e00-b777-6d345d53e3d6","Type":"ContainerStarted","Data":"5c764be6c74ded7275ee948be240c8747a10748063054eebb75b3b6cb45dee93"} Dec 01 16:29:23 crc kubenswrapper[4739]: I1201 16:29:23.722084 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8zzt6" podStartSLOduration=9.127190307 podStartE2EDuration="12.722062585s" podCreationTimestamp="2025-12-01 16:29:11 +0000 UTC" firstStartedPulling="2025-12-01 16:29:19.656912234 +0000 UTC m=+3261.482658368" lastFinishedPulling="2025-12-01 16:29:23.251784552 +0000 UTC m=+3265.077530646" observedRunningTime="2025-12-01 16:29:23.714846451 +0000 UTC m=+3265.540592545" watchObservedRunningTime="2025-12-01 16:29:23.722062585 +0000 UTC m=+3265.547808679" Dec 01 16:29:31 crc kubenswrapper[4739]: I1201 16:29:31.409656 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:31 crc kubenswrapper[4739]: I1201 16:29:31.410681 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:31 crc kubenswrapper[4739]: I1201 16:29:31.468356 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:31 crc kubenswrapper[4739]: I1201 16:29:31.823808 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:31 crc kubenswrapper[4739]: I1201 16:29:31.883769 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8zzt6"] Dec 01 16:29:33 crc kubenswrapper[4739]: I1201 16:29:33.789498 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8zzt6" podUID="fa52dc96-cd9d-4e00-b777-6d345d53e3d6" containerName="registry-server" containerID="cri-o://5c764be6c74ded7275ee948be240c8747a10748063054eebb75b3b6cb45dee93" gracePeriod=2 Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.191376 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.324796 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.490572 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlh7r\" (UniqueName: \"kubernetes.io/projected/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-kube-api-access-xlh7r\") pod \"fa52dc96-cd9d-4e00-b777-6d345d53e3d6\" (UID: \"fa52dc96-cd9d-4e00-b777-6d345d53e3d6\") " Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.490997 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-catalog-content\") pod \"fa52dc96-cd9d-4e00-b777-6d345d53e3d6\" (UID: \"fa52dc96-cd9d-4e00-b777-6d345d53e3d6\") " Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.491144 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-utilities\") pod \"fa52dc96-cd9d-4e00-b777-6d345d53e3d6\" (UID: \"fa52dc96-cd9d-4e00-b777-6d345d53e3d6\") " Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.492135 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-utilities" (OuterVolumeSpecName: "utilities") pod "fa52dc96-cd9d-4e00-b777-6d345d53e3d6" (UID: "fa52dc96-cd9d-4e00-b777-6d345d53e3d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.496614 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-kube-api-access-xlh7r" (OuterVolumeSpecName: "kube-api-access-xlh7r") pod "fa52dc96-cd9d-4e00-b777-6d345d53e3d6" (UID: "fa52dc96-cd9d-4e00-b777-6d345d53e3d6"). InnerVolumeSpecName "kube-api-access-xlh7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.593859 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.593893 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlh7r\" (UniqueName: \"kubernetes.io/projected/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-kube-api-access-xlh7r\") on node \"crc\" DevicePath \"\"" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.604558 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa52dc96-cd9d-4e00-b777-6d345d53e3d6" (UID: "fa52dc96-cd9d-4e00-b777-6d345d53e3d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.696198 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa52dc96-cd9d-4e00-b777-6d345d53e3d6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.802070 4739 generic.go:334] "Generic (PLEG): container finished" podID="fa52dc96-cd9d-4e00-b777-6d345d53e3d6" containerID="5c764be6c74ded7275ee948be240c8747a10748063054eebb75b3b6cb45dee93" exitCode=0 Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.802114 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zzt6" event={"ID":"fa52dc96-cd9d-4e00-b777-6d345d53e3d6","Type":"ContainerDied","Data":"5c764be6c74ded7275ee948be240c8747a10748063054eebb75b3b6cb45dee93"} Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.802140 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zzt6" event={"ID":"fa52dc96-cd9d-4e00-b777-6d345d53e3d6","Type":"ContainerDied","Data":"c28e5d01681df4d36ea18d0cd759c8934039f83ea7a515ee3d72fc539b430f96"} Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.802156 4739 scope.go:117] "RemoveContainer" containerID="5c764be6c74ded7275ee948be240c8747a10748063054eebb75b3b6cb45dee93" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.802269 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zzt6" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.835202 4739 scope.go:117] "RemoveContainer" containerID="0b4f60433ea97f6a6645239b4b128d33d0d368d5a3da6e765b8261bf5d61f280" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.850269 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8zzt6"] Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.857556 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8zzt6"] Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.871602 4739 scope.go:117] "RemoveContainer" containerID="3b5c0da23349c43b19a5fcd17df12430faa26f597de7c54a842a4cfafe4c9b45" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.890728 4739 scope.go:117] "RemoveContainer" containerID="5c764be6c74ded7275ee948be240c8747a10748063054eebb75b3b6cb45dee93" Dec 01 16:29:34 crc kubenswrapper[4739]: E1201 16:29:34.893041 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c764be6c74ded7275ee948be240c8747a10748063054eebb75b3b6cb45dee93\": container with ID starting with 5c764be6c74ded7275ee948be240c8747a10748063054eebb75b3b6cb45dee93 not found: ID does not exist" containerID="5c764be6c74ded7275ee948be240c8747a10748063054eebb75b3b6cb45dee93" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.893101 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c764be6c74ded7275ee948be240c8747a10748063054eebb75b3b6cb45dee93"} err="failed to get container status \"5c764be6c74ded7275ee948be240c8747a10748063054eebb75b3b6cb45dee93\": rpc error: code = NotFound desc = could not find container \"5c764be6c74ded7275ee948be240c8747a10748063054eebb75b3b6cb45dee93\": container with ID starting with 5c764be6c74ded7275ee948be240c8747a10748063054eebb75b3b6cb45dee93 not found: ID does not exist" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.893133 4739 scope.go:117] "RemoveContainer" containerID="0b4f60433ea97f6a6645239b4b128d33d0d368d5a3da6e765b8261bf5d61f280" Dec 01 16:29:34 crc kubenswrapper[4739]: E1201 16:29:34.893644 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b4f60433ea97f6a6645239b4b128d33d0d368d5a3da6e765b8261bf5d61f280\": container with ID starting with 0b4f60433ea97f6a6645239b4b128d33d0d368d5a3da6e765b8261bf5d61f280 not found: ID does not exist" containerID="0b4f60433ea97f6a6645239b4b128d33d0d368d5a3da6e765b8261bf5d61f280" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.893689 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b4f60433ea97f6a6645239b4b128d33d0d368d5a3da6e765b8261bf5d61f280"} err="failed to get container status \"0b4f60433ea97f6a6645239b4b128d33d0d368d5a3da6e765b8261bf5d61f280\": rpc error: code = NotFound desc = could not find container \"0b4f60433ea97f6a6645239b4b128d33d0d368d5a3da6e765b8261bf5d61f280\": container with ID starting with 0b4f60433ea97f6a6645239b4b128d33d0d368d5a3da6e765b8261bf5d61f280 not found: ID does not exist" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.893720 4739 scope.go:117] "RemoveContainer" containerID="3b5c0da23349c43b19a5fcd17df12430faa26f597de7c54a842a4cfafe4c9b45" Dec 01 16:29:34 crc kubenswrapper[4739]: E1201 16:29:34.894027 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b5c0da23349c43b19a5fcd17df12430faa26f597de7c54a842a4cfafe4c9b45\": container with ID starting with 3b5c0da23349c43b19a5fcd17df12430faa26f597de7c54a842a4cfafe4c9b45 not found: ID does not exist" containerID="3b5c0da23349c43b19a5fcd17df12430faa26f597de7c54a842a4cfafe4c9b45" Dec 01 16:29:34 crc kubenswrapper[4739]: I1201 16:29:34.894068 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b5c0da23349c43b19a5fcd17df12430faa26f597de7c54a842a4cfafe4c9b45"} err="failed to get container status \"3b5c0da23349c43b19a5fcd17df12430faa26f597de7c54a842a4cfafe4c9b45\": rpc error: code = NotFound desc = could not find container \"3b5c0da23349c43b19a5fcd17df12430faa26f597de7c54a842a4cfafe4c9b45\": container with ID starting with 3b5c0da23349c43b19a5fcd17df12430faa26f597de7c54a842a4cfafe4c9b45 not found: ID does not exist" Dec 01 16:29:35 crc kubenswrapper[4739]: I1201 16:29:35.819202 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d7438aa0-1e45-4112-9af4-a584825d29df","Type":"ContainerStarted","Data":"cf2aad93d9175d6043838d61f159d79af1a951f700e12bddc32e32a631e77e5f"} Dec 01 16:29:35 crc kubenswrapper[4739]: I1201 16:29:35.854180 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.985661937 podStartE2EDuration="54.854149778s" podCreationTimestamp="2025-12-01 16:28:41 +0000 UTC" firstStartedPulling="2025-12-01 16:28:43.320123139 +0000 UTC m=+3225.145869233" lastFinishedPulling="2025-12-01 16:29:34.18861097 +0000 UTC m=+3276.014357074" observedRunningTime="2025-12-01 16:29:35.844250593 +0000 UTC m=+3277.669996767" watchObservedRunningTime="2025-12-01 16:29:35.854149778 +0000 UTC m=+3277.679895912" Dec 01 16:29:36 crc kubenswrapper[4739]: I1201 16:29:36.492870 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa52dc96-cd9d-4e00-b777-6d345d53e3d6" path="/var/lib/kubelet/pods/fa52dc96-cd9d-4e00-b777-6d345d53e3d6/volumes" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.152617 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl"] Dec 01 16:30:00 crc kubenswrapper[4739]: E1201 16:30:00.154041 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa52dc96-cd9d-4e00-b777-6d345d53e3d6" containerName="extract-utilities" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.154070 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa52dc96-cd9d-4e00-b777-6d345d53e3d6" containerName="extract-utilities" Dec 01 16:30:00 crc kubenswrapper[4739]: E1201 16:30:00.154136 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa52dc96-cd9d-4e00-b777-6d345d53e3d6" containerName="extract-content" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.154150 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa52dc96-cd9d-4e00-b777-6d345d53e3d6" containerName="extract-content" Dec 01 16:30:00 crc kubenswrapper[4739]: E1201 16:30:00.154176 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa52dc96-cd9d-4e00-b777-6d345d53e3d6" containerName="registry-server" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.154190 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa52dc96-cd9d-4e00-b777-6d345d53e3d6" containerName="registry-server" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.154571 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa52dc96-cd9d-4e00-b777-6d345d53e3d6" containerName="registry-server" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.155771 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.159765 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.160506 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.194220 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl"] Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.270944 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf6bd\" (UniqueName: \"kubernetes.io/projected/717822f5-713f-41fe-b1ad-c3699247a6f0-kube-api-access-kf6bd\") pod \"collect-profiles-29410110-kl7rl\" (UID: \"717822f5-713f-41fe-b1ad-c3699247a6f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.271450 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/717822f5-713f-41fe-b1ad-c3699247a6f0-config-volume\") pod \"collect-profiles-29410110-kl7rl\" (UID: \"717822f5-713f-41fe-b1ad-c3699247a6f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.271677 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/717822f5-713f-41fe-b1ad-c3699247a6f0-secret-volume\") pod \"collect-profiles-29410110-kl7rl\" (UID: \"717822f5-713f-41fe-b1ad-c3699247a6f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.373782 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf6bd\" (UniqueName: \"kubernetes.io/projected/717822f5-713f-41fe-b1ad-c3699247a6f0-kube-api-access-kf6bd\") pod \"collect-profiles-29410110-kl7rl\" (UID: \"717822f5-713f-41fe-b1ad-c3699247a6f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.374053 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/717822f5-713f-41fe-b1ad-c3699247a6f0-config-volume\") pod \"collect-profiles-29410110-kl7rl\" (UID: \"717822f5-713f-41fe-b1ad-c3699247a6f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.374184 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/717822f5-713f-41fe-b1ad-c3699247a6f0-secret-volume\") pod \"collect-profiles-29410110-kl7rl\" (UID: \"717822f5-713f-41fe-b1ad-c3699247a6f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.375650 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/717822f5-713f-41fe-b1ad-c3699247a6f0-config-volume\") pod \"collect-profiles-29410110-kl7rl\" (UID: \"717822f5-713f-41fe-b1ad-c3699247a6f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.389058 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/717822f5-713f-41fe-b1ad-c3699247a6f0-secret-volume\") pod \"collect-profiles-29410110-kl7rl\" (UID: \"717822f5-713f-41fe-b1ad-c3699247a6f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.397341 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf6bd\" (UniqueName: \"kubernetes.io/projected/717822f5-713f-41fe-b1ad-c3699247a6f0-kube-api-access-kf6bd\") pod \"collect-profiles-29410110-kl7rl\" (UID: \"717822f5-713f-41fe-b1ad-c3699247a6f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" Dec 01 16:30:00 crc kubenswrapper[4739]: I1201 16:30:00.487745 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" Dec 01 16:30:01 crc kubenswrapper[4739]: I1201 16:30:01.045914 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl"] Dec 01 16:30:01 crc kubenswrapper[4739]: I1201 16:30:01.206649 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" event={"ID":"717822f5-713f-41fe-b1ad-c3699247a6f0","Type":"ContainerStarted","Data":"b55667548961a284803c82b71f038aea34880932d205269c4483130a3dc53349"} Dec 01 16:30:02 crc kubenswrapper[4739]: I1201 16:30:02.221302 4739 generic.go:334] "Generic (PLEG): container finished" podID="717822f5-713f-41fe-b1ad-c3699247a6f0" containerID="e8b3d60c77e3794126363900e351738626ff621a5bf8eaaa51b83bc88980d4b7" exitCode=0 Dec 01 16:30:02 crc kubenswrapper[4739]: I1201 16:30:02.221385 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" event={"ID":"717822f5-713f-41fe-b1ad-c3699247a6f0","Type":"ContainerDied","Data":"e8b3d60c77e3794126363900e351738626ff621a5bf8eaaa51b83bc88980d4b7"} Dec 01 16:30:03 crc kubenswrapper[4739]: I1201 16:30:03.591356 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" Dec 01 16:30:03 crc kubenswrapper[4739]: I1201 16:30:03.639406 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/717822f5-713f-41fe-b1ad-c3699247a6f0-secret-volume\") pod \"717822f5-713f-41fe-b1ad-c3699247a6f0\" (UID: \"717822f5-713f-41fe-b1ad-c3699247a6f0\") " Dec 01 16:30:03 crc kubenswrapper[4739]: I1201 16:30:03.639545 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kf6bd\" (UniqueName: \"kubernetes.io/projected/717822f5-713f-41fe-b1ad-c3699247a6f0-kube-api-access-kf6bd\") pod \"717822f5-713f-41fe-b1ad-c3699247a6f0\" (UID: \"717822f5-713f-41fe-b1ad-c3699247a6f0\") " Dec 01 16:30:03 crc kubenswrapper[4739]: I1201 16:30:03.639647 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/717822f5-713f-41fe-b1ad-c3699247a6f0-config-volume\") pod \"717822f5-713f-41fe-b1ad-c3699247a6f0\" (UID: \"717822f5-713f-41fe-b1ad-c3699247a6f0\") " Dec 01 16:30:03 crc kubenswrapper[4739]: I1201 16:30:03.640605 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/717822f5-713f-41fe-b1ad-c3699247a6f0-config-volume" (OuterVolumeSpecName: "config-volume") pod "717822f5-713f-41fe-b1ad-c3699247a6f0" (UID: "717822f5-713f-41fe-b1ad-c3699247a6f0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:30:03 crc kubenswrapper[4739]: I1201 16:30:03.649509 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/717822f5-713f-41fe-b1ad-c3699247a6f0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "717822f5-713f-41fe-b1ad-c3699247a6f0" (UID: "717822f5-713f-41fe-b1ad-c3699247a6f0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:30:03 crc kubenswrapper[4739]: I1201 16:30:03.649604 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/717822f5-713f-41fe-b1ad-c3699247a6f0-kube-api-access-kf6bd" (OuterVolumeSpecName: "kube-api-access-kf6bd") pod "717822f5-713f-41fe-b1ad-c3699247a6f0" (UID: "717822f5-713f-41fe-b1ad-c3699247a6f0"). InnerVolumeSpecName "kube-api-access-kf6bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:30:03 crc kubenswrapper[4739]: I1201 16:30:03.741960 4739 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/717822f5-713f-41fe-b1ad-c3699247a6f0-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:30:03 crc kubenswrapper[4739]: I1201 16:30:03.742001 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kf6bd\" (UniqueName: \"kubernetes.io/projected/717822f5-713f-41fe-b1ad-c3699247a6f0-kube-api-access-kf6bd\") on node \"crc\" DevicePath \"\"" Dec 01 16:30:03 crc kubenswrapper[4739]: I1201 16:30:03.742011 4739 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/717822f5-713f-41fe-b1ad-c3699247a6f0-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:30:04 crc kubenswrapper[4739]: I1201 16:30:04.245307 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" event={"ID":"717822f5-713f-41fe-b1ad-c3699247a6f0","Type":"ContainerDied","Data":"b55667548961a284803c82b71f038aea34880932d205269c4483130a3dc53349"} Dec 01 16:30:04 crc kubenswrapper[4739]: I1201 16:30:04.245372 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b55667548961a284803c82b71f038aea34880932d205269c4483130a3dc53349" Dec 01 16:30:04 crc kubenswrapper[4739]: I1201 16:30:04.245375 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-kl7rl" Dec 01 16:30:04 crc kubenswrapper[4739]: I1201 16:30:04.687553 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5"] Dec 01 16:30:04 crc kubenswrapper[4739]: I1201 16:30:04.698826 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410065-nq4j5"] Dec 01 16:30:06 crc kubenswrapper[4739]: I1201 16:30:06.499727 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6950571-e8f3-4e9c-b2e8-05a806722f8b" path="/var/lib/kubelet/pods/e6950571-e8f3-4e9c-b2e8-05a806722f8b/volumes" Dec 01 16:30:09 crc kubenswrapper[4739]: I1201 16:30:09.622656 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:30:09 crc kubenswrapper[4739]: I1201 16:30:09.623414 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:30:24 crc kubenswrapper[4739]: I1201 16:30:24.403440 4739 scope.go:117] "RemoveContainer" containerID="11fa0ea5b5cf8b38a4664f369c5f50fb93ddfa0697829736c99ba7bdc953acef" Dec 01 16:30:39 crc kubenswrapper[4739]: I1201 16:30:39.622399 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:30:39 crc kubenswrapper[4739]: I1201 16:30:39.623174 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:31:09 crc kubenswrapper[4739]: I1201 16:31:09.621978 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:31:09 crc kubenswrapper[4739]: I1201 16:31:09.622886 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:31:09 crc kubenswrapper[4739]: I1201 16:31:09.622962 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 16:31:09 crc kubenswrapper[4739]: I1201 16:31:09.624287 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8c69d57ff87a87518f586ab31047e56d7736ae4a7c0a6aae691054d7b0c41c0d"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:31:09 crc kubenswrapper[4739]: I1201 16:31:09.624404 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://8c69d57ff87a87518f586ab31047e56d7736ae4a7c0a6aae691054d7b0c41c0d" gracePeriod=600 Dec 01 16:31:10 crc kubenswrapper[4739]: I1201 16:31:10.061721 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="8c69d57ff87a87518f586ab31047e56d7736ae4a7c0a6aae691054d7b0c41c0d" exitCode=0 Dec 01 16:31:10 crc kubenswrapper[4739]: I1201 16:31:10.061817 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"8c69d57ff87a87518f586ab31047e56d7736ae4a7c0a6aae691054d7b0c41c0d"} Dec 01 16:31:10 crc kubenswrapper[4739]: I1201 16:31:10.062817 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364"} Dec 01 16:31:10 crc kubenswrapper[4739]: I1201 16:31:10.062892 4739 scope.go:117] "RemoveContainer" containerID="b4906dd73338c5f45751d988a864bc60981bf9b38ce1520742b03698cbb4d83d" Dec 01 16:33:09 crc kubenswrapper[4739]: I1201 16:33:09.622405 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:33:09 crc kubenswrapper[4739]: I1201 16:33:09.623004 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.468920 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r4zlg"] Dec 01 16:33:21 crc kubenswrapper[4739]: E1201 16:33:21.470683 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="717822f5-713f-41fe-b1ad-c3699247a6f0" containerName="collect-profiles" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.470760 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="717822f5-713f-41fe-b1ad-c3699247a6f0" containerName="collect-profiles" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.471003 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="717822f5-713f-41fe-b1ad-c3699247a6f0" containerName="collect-profiles" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.472532 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.485276 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r4zlg"] Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.646495 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqhth\" (UniqueName: \"kubernetes.io/projected/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-kube-api-access-dqhth\") pod \"redhat-marketplace-r4zlg\" (UID: \"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1\") " pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.646670 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-utilities\") pod \"redhat-marketplace-r4zlg\" (UID: \"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1\") " pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.646730 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-catalog-content\") pod \"redhat-marketplace-r4zlg\" (UID: \"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1\") " pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.672907 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xq8q7"] Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.675266 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xq8q7" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.685831 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xq8q7"] Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.748724 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-utilities\") pod \"redhat-marketplace-r4zlg\" (UID: \"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1\") " pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.748789 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-catalog-content\") pod \"redhat-marketplace-r4zlg\" (UID: \"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1\") " pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.748855 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqhth\" (UniqueName: \"kubernetes.io/projected/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-kube-api-access-dqhth\") pod \"redhat-marketplace-r4zlg\" (UID: \"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1\") " pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.749292 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-utilities\") pod \"redhat-marketplace-r4zlg\" (UID: \"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1\") " pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.749568 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-catalog-content\") pod \"redhat-marketplace-r4zlg\" (UID: \"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1\") " pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.770267 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqhth\" (UniqueName: \"kubernetes.io/projected/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-kube-api-access-dqhth\") pod \"redhat-marketplace-r4zlg\" (UID: \"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1\") " pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.804264 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.850762 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r45bs\" (UniqueName: \"kubernetes.io/projected/d5b77aa5-36d6-4cfd-8742-e71853ba16dd-kube-api-access-r45bs\") pod \"certified-operators-xq8q7\" (UID: \"d5b77aa5-36d6-4cfd-8742-e71853ba16dd\") " pod="openshift-marketplace/certified-operators-xq8q7" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.851223 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5b77aa5-36d6-4cfd-8742-e71853ba16dd-utilities\") pod \"certified-operators-xq8q7\" (UID: \"d5b77aa5-36d6-4cfd-8742-e71853ba16dd\") " pod="openshift-marketplace/certified-operators-xq8q7" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.851407 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5b77aa5-36d6-4cfd-8742-e71853ba16dd-catalog-content\") pod \"certified-operators-xq8q7\" (UID: \"d5b77aa5-36d6-4cfd-8742-e71853ba16dd\") " pod="openshift-marketplace/certified-operators-xq8q7" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.953301 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r45bs\" (UniqueName: \"kubernetes.io/projected/d5b77aa5-36d6-4cfd-8742-e71853ba16dd-kube-api-access-r45bs\") pod \"certified-operators-xq8q7\" (UID: \"d5b77aa5-36d6-4cfd-8742-e71853ba16dd\") " pod="openshift-marketplace/certified-operators-xq8q7" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.953352 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5b77aa5-36d6-4cfd-8742-e71853ba16dd-utilities\") pod \"certified-operators-xq8q7\" (UID: \"d5b77aa5-36d6-4cfd-8742-e71853ba16dd\") " pod="openshift-marketplace/certified-operators-xq8q7" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.953461 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5b77aa5-36d6-4cfd-8742-e71853ba16dd-catalog-content\") pod \"certified-operators-xq8q7\" (UID: \"d5b77aa5-36d6-4cfd-8742-e71853ba16dd\") " pod="openshift-marketplace/certified-operators-xq8q7" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.957588 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5b77aa5-36d6-4cfd-8742-e71853ba16dd-catalog-content\") pod \"certified-operators-xq8q7\" (UID: \"d5b77aa5-36d6-4cfd-8742-e71853ba16dd\") " pod="openshift-marketplace/certified-operators-xq8q7" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.961741 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5b77aa5-36d6-4cfd-8742-e71853ba16dd-utilities\") pod \"certified-operators-xq8q7\" (UID: \"d5b77aa5-36d6-4cfd-8742-e71853ba16dd\") " pod="openshift-marketplace/certified-operators-xq8q7" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.986206 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r45bs\" (UniqueName: \"kubernetes.io/projected/d5b77aa5-36d6-4cfd-8742-e71853ba16dd-kube-api-access-r45bs\") pod \"certified-operators-xq8q7\" (UID: \"d5b77aa5-36d6-4cfd-8742-e71853ba16dd\") " pod="openshift-marketplace/certified-operators-xq8q7" Dec 01 16:33:21 crc kubenswrapper[4739]: I1201 16:33:21.992099 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xq8q7" Dec 01 16:33:22 crc kubenswrapper[4739]: I1201 16:33:22.412221 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r4zlg"] Dec 01 16:33:22 crc kubenswrapper[4739]: I1201 16:33:22.515093 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xq8q7"] Dec 01 16:33:22 crc kubenswrapper[4739]: W1201 16:33:22.524016 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5b77aa5_36d6_4cfd_8742_e71853ba16dd.slice/crio-7bea0b77be4756fa00aaaa36858171728c2c42d5ef43ca8f0e99294b7d5af420 WatchSource:0}: Error finding container 7bea0b77be4756fa00aaaa36858171728c2c42d5ef43ca8f0e99294b7d5af420: Status 404 returned error can't find the container with id 7bea0b77be4756fa00aaaa36858171728c2c42d5ef43ca8f0e99294b7d5af420 Dec 01 16:33:22 crc kubenswrapper[4739]: I1201 16:33:22.553254 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r4zlg" event={"ID":"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1","Type":"ContainerStarted","Data":"92d85e56a9095d779c69e84f1c777d878bf1873124c650510aee6934bf238c44"} Dec 01 16:33:22 crc kubenswrapper[4739]: I1201 16:33:22.554671 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xq8q7" event={"ID":"d5b77aa5-36d6-4cfd-8742-e71853ba16dd","Type":"ContainerStarted","Data":"7bea0b77be4756fa00aaaa36858171728c2c42d5ef43ca8f0e99294b7d5af420"} Dec 01 16:33:23 crc kubenswrapper[4739]: I1201 16:33:23.563860 4739 generic.go:334] "Generic (PLEG): container finished" podID="fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1" containerID="d0c2e31beb055521b73855a33859da3ad4c66b2f599f918958bea602ec558384" exitCode=0 Dec 01 16:33:23 crc kubenswrapper[4739]: I1201 16:33:23.563915 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r4zlg" event={"ID":"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1","Type":"ContainerDied","Data":"d0c2e31beb055521b73855a33859da3ad4c66b2f599f918958bea602ec558384"} Dec 01 16:33:23 crc kubenswrapper[4739]: I1201 16:33:23.566559 4739 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:33:23 crc kubenswrapper[4739]: I1201 16:33:23.566642 4739 generic.go:334] "Generic (PLEG): container finished" podID="d5b77aa5-36d6-4cfd-8742-e71853ba16dd" containerID="c3640c0d4affca98fc04809daf71ac8becf279e27d1f009f05055a6599e77d0d" exitCode=0 Dec 01 16:33:23 crc kubenswrapper[4739]: I1201 16:33:23.566686 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xq8q7" event={"ID":"d5b77aa5-36d6-4cfd-8742-e71853ba16dd","Type":"ContainerDied","Data":"c3640c0d4affca98fc04809daf71ac8becf279e27d1f009f05055a6599e77d0d"} Dec 01 16:33:24 crc kubenswrapper[4739]: I1201 16:33:24.571496 4739 scope.go:117] "RemoveContainer" containerID="15897500412d62529446a37ce315bde647579396696d1eec233934952bde65a1" Dec 01 16:33:24 crc kubenswrapper[4739]: I1201 16:33:24.625556 4739 scope.go:117] "RemoveContainer" containerID="5df9727619fac02426e2b001a1d2dd68d53964ca1359dcd3f7b83959b9b4be85" Dec 01 16:33:25 crc kubenswrapper[4739]: I1201 16:33:25.594025 4739 generic.go:334] "Generic (PLEG): container finished" podID="fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1" containerID="012078c3e2dd2844aaea71c4e1ecea82413cedcafa451c056d6840f74edbff82" exitCode=0 Dec 01 16:33:25 crc kubenswrapper[4739]: I1201 16:33:25.594070 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r4zlg" event={"ID":"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1","Type":"ContainerDied","Data":"012078c3e2dd2844aaea71c4e1ecea82413cedcafa451c056d6840f74edbff82"} Dec 01 16:33:26 crc kubenswrapper[4739]: I1201 16:33:26.606765 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r4zlg" event={"ID":"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1","Type":"ContainerStarted","Data":"11d9e0d610b67a7b3862fd77b982a4b5aa050e6b8be05e6b65fa33c73c167d95"} Dec 01 16:33:26 crc kubenswrapper[4739]: I1201 16:33:26.635699 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r4zlg" podStartSLOduration=2.917677679 podStartE2EDuration="5.635676427s" podCreationTimestamp="2025-12-01 16:33:21 +0000 UTC" firstStartedPulling="2025-12-01 16:33:23.566227063 +0000 UTC m=+3505.391973157" lastFinishedPulling="2025-12-01 16:33:26.284225811 +0000 UTC m=+3508.109971905" observedRunningTime="2025-12-01 16:33:26.625723749 +0000 UTC m=+3508.451469843" watchObservedRunningTime="2025-12-01 16:33:26.635676427 +0000 UTC m=+3508.461422521" Dec 01 16:33:29 crc kubenswrapper[4739]: I1201 16:33:29.647071 4739 generic.go:334] "Generic (PLEG): container finished" podID="d5b77aa5-36d6-4cfd-8742-e71853ba16dd" containerID="8f6f99dd5ef409e08947f4e8a65c21dec16b46968ee262559ca3e497d9631a5b" exitCode=0 Dec 01 16:33:29 crc kubenswrapper[4739]: I1201 16:33:29.647181 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xq8q7" event={"ID":"d5b77aa5-36d6-4cfd-8742-e71853ba16dd","Type":"ContainerDied","Data":"8f6f99dd5ef409e08947f4e8a65c21dec16b46968ee262559ca3e497d9631a5b"} Dec 01 16:33:31 crc kubenswrapper[4739]: I1201 16:33:31.673350 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xq8q7" event={"ID":"d5b77aa5-36d6-4cfd-8742-e71853ba16dd","Type":"ContainerStarted","Data":"430b01650e8d1b07d7c46dd86e219684305c0b5a1342215fffaa4bffacc22aeb"} Dec 01 16:33:31 crc kubenswrapper[4739]: I1201 16:33:31.805454 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:31 crc kubenswrapper[4739]: I1201 16:33:31.805497 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:31 crc kubenswrapper[4739]: I1201 16:33:31.870132 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:31 crc kubenswrapper[4739]: I1201 16:33:31.887980 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xq8q7" podStartSLOduration=4.006126893 podStartE2EDuration="10.887965599s" podCreationTimestamp="2025-12-01 16:33:21 +0000 UTC" firstStartedPulling="2025-12-01 16:33:23.567831704 +0000 UTC m=+3505.393577798" lastFinishedPulling="2025-12-01 16:33:30.4496704 +0000 UTC m=+3512.275416504" observedRunningTime="2025-12-01 16:33:31.696884346 +0000 UTC m=+3513.522630440" watchObservedRunningTime="2025-12-01 16:33:31.887965599 +0000 UTC m=+3513.713711693" Dec 01 16:33:31 crc kubenswrapper[4739]: I1201 16:33:31.992774 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xq8q7" Dec 01 16:33:31 crc kubenswrapper[4739]: I1201 16:33:31.993202 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xq8q7" Dec 01 16:33:32 crc kubenswrapper[4739]: I1201 16:33:32.730633 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:33 crc kubenswrapper[4739]: I1201 16:33:33.050325 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-xq8q7" podUID="d5b77aa5-36d6-4cfd-8742-e71853ba16dd" containerName="registry-server" probeResult="failure" output=< Dec 01 16:33:33 crc kubenswrapper[4739]: timeout: failed to connect service ":50051" within 1s Dec 01 16:33:33 crc kubenswrapper[4739]: > Dec 01 16:33:35 crc kubenswrapper[4739]: I1201 16:33:35.670560 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r4zlg"] Dec 01 16:33:35 crc kubenswrapper[4739]: I1201 16:33:35.671191 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-r4zlg" podUID="fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1" containerName="registry-server" containerID="cri-o://11d9e0d610b67a7b3862fd77b982a4b5aa050e6b8be05e6b65fa33c73c167d95" gracePeriod=2 Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.233082 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.294741 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-catalog-content\") pod \"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1\" (UID: \"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1\") " Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.294825 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqhth\" (UniqueName: \"kubernetes.io/projected/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-kube-api-access-dqhth\") pod \"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1\" (UID: \"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1\") " Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.294863 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-utilities\") pod \"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1\" (UID: \"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1\") " Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.296153 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-utilities" (OuterVolumeSpecName: "utilities") pod "fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1" (UID: "fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.322970 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1" (UID: "fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.324689 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-kube-api-access-dqhth" (OuterVolumeSpecName: "kube-api-access-dqhth") pod "fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1" (UID: "fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1"). InnerVolumeSpecName "kube-api-access-dqhth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.396882 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqhth\" (UniqueName: \"kubernetes.io/projected/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-kube-api-access-dqhth\") on node \"crc\" DevicePath \"\"" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.397083 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.397166 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.724027 4739 generic.go:334] "Generic (PLEG): container finished" podID="fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1" containerID="11d9e0d610b67a7b3862fd77b982a4b5aa050e6b8be05e6b65fa33c73c167d95" exitCode=0 Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.724100 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r4zlg" event={"ID":"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1","Type":"ContainerDied","Data":"11d9e0d610b67a7b3862fd77b982a4b5aa050e6b8be05e6b65fa33c73c167d95"} Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.724136 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r4zlg" event={"ID":"fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1","Type":"ContainerDied","Data":"92d85e56a9095d779c69e84f1c777d878bf1873124c650510aee6934bf238c44"} Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.724179 4739 scope.go:117] "RemoveContainer" containerID="11d9e0d610b67a7b3862fd77b982a4b5aa050e6b8be05e6b65fa33c73c167d95" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.724859 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r4zlg" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.753372 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r4zlg"] Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.757725 4739 scope.go:117] "RemoveContainer" containerID="012078c3e2dd2844aaea71c4e1ecea82413cedcafa451c056d6840f74edbff82" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.763141 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-r4zlg"] Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.779224 4739 scope.go:117] "RemoveContainer" containerID="d0c2e31beb055521b73855a33859da3ad4c66b2f599f918958bea602ec558384" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.842472 4739 scope.go:117] "RemoveContainer" containerID="11d9e0d610b67a7b3862fd77b982a4b5aa050e6b8be05e6b65fa33c73c167d95" Dec 01 16:33:36 crc kubenswrapper[4739]: E1201 16:33:36.843216 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11d9e0d610b67a7b3862fd77b982a4b5aa050e6b8be05e6b65fa33c73c167d95\": container with ID starting with 11d9e0d610b67a7b3862fd77b982a4b5aa050e6b8be05e6b65fa33c73c167d95 not found: ID does not exist" containerID="11d9e0d610b67a7b3862fd77b982a4b5aa050e6b8be05e6b65fa33c73c167d95" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.843260 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11d9e0d610b67a7b3862fd77b982a4b5aa050e6b8be05e6b65fa33c73c167d95"} err="failed to get container status \"11d9e0d610b67a7b3862fd77b982a4b5aa050e6b8be05e6b65fa33c73c167d95\": rpc error: code = NotFound desc = could not find container \"11d9e0d610b67a7b3862fd77b982a4b5aa050e6b8be05e6b65fa33c73c167d95\": container with ID starting with 11d9e0d610b67a7b3862fd77b982a4b5aa050e6b8be05e6b65fa33c73c167d95 not found: ID does not exist" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.843289 4739 scope.go:117] "RemoveContainer" containerID="012078c3e2dd2844aaea71c4e1ecea82413cedcafa451c056d6840f74edbff82" Dec 01 16:33:36 crc kubenswrapper[4739]: E1201 16:33:36.844211 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"012078c3e2dd2844aaea71c4e1ecea82413cedcafa451c056d6840f74edbff82\": container with ID starting with 012078c3e2dd2844aaea71c4e1ecea82413cedcafa451c056d6840f74edbff82 not found: ID does not exist" containerID="012078c3e2dd2844aaea71c4e1ecea82413cedcafa451c056d6840f74edbff82" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.844248 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"012078c3e2dd2844aaea71c4e1ecea82413cedcafa451c056d6840f74edbff82"} err="failed to get container status \"012078c3e2dd2844aaea71c4e1ecea82413cedcafa451c056d6840f74edbff82\": rpc error: code = NotFound desc = could not find container \"012078c3e2dd2844aaea71c4e1ecea82413cedcafa451c056d6840f74edbff82\": container with ID starting with 012078c3e2dd2844aaea71c4e1ecea82413cedcafa451c056d6840f74edbff82 not found: ID does not exist" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.844274 4739 scope.go:117] "RemoveContainer" containerID="d0c2e31beb055521b73855a33859da3ad4c66b2f599f918958bea602ec558384" Dec 01 16:33:36 crc kubenswrapper[4739]: E1201 16:33:36.844673 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0c2e31beb055521b73855a33859da3ad4c66b2f599f918958bea602ec558384\": container with ID starting with d0c2e31beb055521b73855a33859da3ad4c66b2f599f918958bea602ec558384 not found: ID does not exist" containerID="d0c2e31beb055521b73855a33859da3ad4c66b2f599f918958bea602ec558384" Dec 01 16:33:36 crc kubenswrapper[4739]: I1201 16:33:36.844757 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0c2e31beb055521b73855a33859da3ad4c66b2f599f918958bea602ec558384"} err="failed to get container status \"d0c2e31beb055521b73855a33859da3ad4c66b2f599f918958bea602ec558384\": rpc error: code = NotFound desc = could not find container \"d0c2e31beb055521b73855a33859da3ad4c66b2f599f918958bea602ec558384\": container with ID starting with d0c2e31beb055521b73855a33859da3ad4c66b2f599f918958bea602ec558384 not found: ID does not exist" Dec 01 16:33:38 crc kubenswrapper[4739]: I1201 16:33:38.490766 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1" path="/var/lib/kubelet/pods/fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1/volumes" Dec 01 16:33:39 crc kubenswrapper[4739]: I1201 16:33:39.621518 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:33:39 crc kubenswrapper[4739]: I1201 16:33:39.621604 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:33:42 crc kubenswrapper[4739]: I1201 16:33:42.046184 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xq8q7" Dec 01 16:33:42 crc kubenswrapper[4739]: I1201 16:33:42.111513 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xq8q7" Dec 01 16:33:42 crc kubenswrapper[4739]: I1201 16:33:42.180908 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xq8q7"] Dec 01 16:33:42 crc kubenswrapper[4739]: I1201 16:33:42.286767 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w7nfj"] Dec 01 16:33:42 crc kubenswrapper[4739]: I1201 16:33:42.287010 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w7nfj" podUID="875e9c09-a5e1-455a-b12e-e46ea43be9ed" containerName="registry-server" containerID="cri-o://9c2dcb5020e4c998e852d33aee7b8d35efeba821df91601ea0cf7d51d8c681f3" gracePeriod=2 Dec 01 16:33:42 crc kubenswrapper[4739]: I1201 16:33:42.785066 4739 generic.go:334] "Generic (PLEG): container finished" podID="875e9c09-a5e1-455a-b12e-e46ea43be9ed" containerID="9c2dcb5020e4c998e852d33aee7b8d35efeba821df91601ea0cf7d51d8c681f3" exitCode=0 Dec 01 16:33:42 crc kubenswrapper[4739]: I1201 16:33:42.785509 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7nfj" event={"ID":"875e9c09-a5e1-455a-b12e-e46ea43be9ed","Type":"ContainerDied","Data":"9c2dcb5020e4c998e852d33aee7b8d35efeba821df91601ea0cf7d51d8c681f3"} Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.096258 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.146565 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq8g9\" (UniqueName: \"kubernetes.io/projected/875e9c09-a5e1-455a-b12e-e46ea43be9ed-kube-api-access-tq8g9\") pod \"875e9c09-a5e1-455a-b12e-e46ea43be9ed\" (UID: \"875e9c09-a5e1-455a-b12e-e46ea43be9ed\") " Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.146725 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/875e9c09-a5e1-455a-b12e-e46ea43be9ed-utilities\") pod \"875e9c09-a5e1-455a-b12e-e46ea43be9ed\" (UID: \"875e9c09-a5e1-455a-b12e-e46ea43be9ed\") " Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.146862 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/875e9c09-a5e1-455a-b12e-e46ea43be9ed-catalog-content\") pod \"875e9c09-a5e1-455a-b12e-e46ea43be9ed\" (UID: \"875e9c09-a5e1-455a-b12e-e46ea43be9ed\") " Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.156923 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/875e9c09-a5e1-455a-b12e-e46ea43be9ed-utilities" (OuterVolumeSpecName: "utilities") pod "875e9c09-a5e1-455a-b12e-e46ea43be9ed" (UID: "875e9c09-a5e1-455a-b12e-e46ea43be9ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.169963 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/875e9c09-a5e1-455a-b12e-e46ea43be9ed-kube-api-access-tq8g9" (OuterVolumeSpecName: "kube-api-access-tq8g9") pod "875e9c09-a5e1-455a-b12e-e46ea43be9ed" (UID: "875e9c09-a5e1-455a-b12e-e46ea43be9ed"). InnerVolumeSpecName "kube-api-access-tq8g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.203145 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/875e9c09-a5e1-455a-b12e-e46ea43be9ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "875e9c09-a5e1-455a-b12e-e46ea43be9ed" (UID: "875e9c09-a5e1-455a-b12e-e46ea43be9ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.248821 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/875e9c09-a5e1-455a-b12e-e46ea43be9ed-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.249146 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/875e9c09-a5e1-455a-b12e-e46ea43be9ed-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.249162 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq8g9\" (UniqueName: \"kubernetes.io/projected/875e9c09-a5e1-455a-b12e-e46ea43be9ed-kube-api-access-tq8g9\") on node \"crc\" DevicePath \"\"" Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.797624 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7nfj" event={"ID":"875e9c09-a5e1-455a-b12e-e46ea43be9ed","Type":"ContainerDied","Data":"dac444c5593671b13f79c430bb7495aee4ac6fd412a0cc72d697af59ff0343f4"} Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.797691 4739 scope.go:117] "RemoveContainer" containerID="9c2dcb5020e4c998e852d33aee7b8d35efeba821df91601ea0cf7d51d8c681f3" Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.797690 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w7nfj" Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.841670 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w7nfj"] Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.843945 4739 scope.go:117] "RemoveContainer" containerID="452fbb2768d590796077dca8baefba28c0ff7e9cf3ec5c21f54de086bbba6070" Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.852229 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w7nfj"] Dec 01 16:33:43 crc kubenswrapper[4739]: I1201 16:33:43.876533 4739 scope.go:117] "RemoveContainer" containerID="4518238ed1c7272568388d3699d0f938dca110865554d0eada543bbe739f859a" Dec 01 16:33:44 crc kubenswrapper[4739]: I1201 16:33:44.488120 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="875e9c09-a5e1-455a-b12e-e46ea43be9ed" path="/var/lib/kubelet/pods/875e9c09-a5e1-455a-b12e-e46ea43be9ed/volumes" Dec 01 16:34:09 crc kubenswrapper[4739]: I1201 16:34:09.621604 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:34:09 crc kubenswrapper[4739]: I1201 16:34:09.622055 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:34:09 crc kubenswrapper[4739]: I1201 16:34:09.622095 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 16:34:09 crc kubenswrapper[4739]: I1201 16:34:09.622608 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:34:09 crc kubenswrapper[4739]: I1201 16:34:09.622654 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" gracePeriod=600 Dec 01 16:34:09 crc kubenswrapper[4739]: E1201 16:34:09.754530 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:34:10 crc kubenswrapper[4739]: I1201 16:34:10.049482 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" exitCode=0 Dec 01 16:34:10 crc kubenswrapper[4739]: I1201 16:34:10.049541 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364"} Dec 01 16:34:10 crc kubenswrapper[4739]: I1201 16:34:10.049595 4739 scope.go:117] "RemoveContainer" containerID="8c69d57ff87a87518f586ab31047e56d7736ae4a7c0a6aae691054d7b0c41c0d" Dec 01 16:34:10 crc kubenswrapper[4739]: I1201 16:34:10.050452 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:34:10 crc kubenswrapper[4739]: E1201 16:34:10.050781 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:34:21 crc kubenswrapper[4739]: I1201 16:34:21.477481 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:34:21 crc kubenswrapper[4739]: E1201 16:34:21.478536 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:34:32 crc kubenswrapper[4739]: I1201 16:34:32.477970 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:34:32 crc kubenswrapper[4739]: E1201 16:34:32.479064 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:34:45 crc kubenswrapper[4739]: I1201 16:34:45.477287 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:34:45 crc kubenswrapper[4739]: E1201 16:34:45.478493 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:34:56 crc kubenswrapper[4739]: I1201 16:34:56.477681 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:34:56 crc kubenswrapper[4739]: E1201 16:34:56.478658 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:35:10 crc kubenswrapper[4739]: I1201 16:35:10.477305 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:35:10 crc kubenswrapper[4739]: E1201 16:35:10.478114 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:35:23 crc kubenswrapper[4739]: I1201 16:35:23.477341 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:35:23 crc kubenswrapper[4739]: E1201 16:35:23.478069 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:35:37 crc kubenswrapper[4739]: I1201 16:35:37.487707 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:35:37 crc kubenswrapper[4739]: E1201 16:35:37.489301 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:35:52 crc kubenswrapper[4739]: I1201 16:35:52.477527 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:35:52 crc kubenswrapper[4739]: E1201 16:35:52.478623 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:36:04 crc kubenswrapper[4739]: I1201 16:36:04.478065 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:36:04 crc kubenswrapper[4739]: E1201 16:36:04.479352 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:36:15 crc kubenswrapper[4739]: I1201 16:36:15.477865 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:36:15 crc kubenswrapper[4739]: E1201 16:36:15.479056 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:36:28 crc kubenswrapper[4739]: I1201 16:36:28.502408 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:36:28 crc kubenswrapper[4739]: E1201 16:36:28.503820 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:36:30 crc kubenswrapper[4739]: I1201 16:36:30.052069 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-28srx"] Dec 01 16:36:30 crc kubenswrapper[4739]: I1201 16:36:30.064604 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-29c8-account-create-update-nd9g7"] Dec 01 16:36:30 crc kubenswrapper[4739]: I1201 16:36:30.075001 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-28srx"] Dec 01 16:36:30 crc kubenswrapper[4739]: I1201 16:36:30.092060 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-29c8-account-create-update-nd9g7"] Dec 01 16:36:30 crc kubenswrapper[4739]: I1201 16:36:30.488932 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="321be3c7-476a-4d82-8bde-4a5d54f5ab9d" path="/var/lib/kubelet/pods/321be3c7-476a-4d82-8bde-4a5d54f5ab9d/volumes" Dec 01 16:36:30 crc kubenswrapper[4739]: I1201 16:36:30.490642 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9a7eb0a-9824-40c2-b44f-2ce4deb7604e" path="/var/lib/kubelet/pods/d9a7eb0a-9824-40c2-b44f-2ce4deb7604e/volumes" Dec 01 16:36:42 crc kubenswrapper[4739]: I1201 16:36:42.478034 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:36:42 crc kubenswrapper[4739]: E1201 16:36:42.479121 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:36:54 crc kubenswrapper[4739]: I1201 16:36:54.477266 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:36:54 crc kubenswrapper[4739]: E1201 16:36:54.478519 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:36:55 crc kubenswrapper[4739]: I1201 16:36:55.073736 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-qkq9f"] Dec 01 16:36:55 crc kubenswrapper[4739]: I1201 16:36:55.081990 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-qkq9f"] Dec 01 16:36:56 crc kubenswrapper[4739]: I1201 16:36:56.491415 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca6268bc-3fab-4df5-a86e-f2060ab9abb8" path="/var/lib/kubelet/pods/ca6268bc-3fab-4df5-a86e-f2060ab9abb8/volumes" Dec 01 16:37:06 crc kubenswrapper[4739]: I1201 16:37:06.477195 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:37:06 crc kubenswrapper[4739]: E1201 16:37:06.478029 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:37:17 crc kubenswrapper[4739]: I1201 16:37:17.478132 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:37:17 crc kubenswrapper[4739]: E1201 16:37:17.479074 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:37:24 crc kubenswrapper[4739]: I1201 16:37:24.777347 4739 scope.go:117] "RemoveContainer" containerID="db11f88319bf52fb097b2224554764287f16f05f3e94485dad5a65ede39487f4" Dec 01 16:37:24 crc kubenswrapper[4739]: I1201 16:37:24.807529 4739 scope.go:117] "RemoveContainer" containerID="a1850cb88b9bec4f31329fada90766c4c363892ffb306b9eae14c4f2e986eb4a" Dec 01 16:37:24 crc kubenswrapper[4739]: I1201 16:37:24.847575 4739 scope.go:117] "RemoveContainer" containerID="c5797f92b4f51634e7a03ff937825a6260c67821d162a5200ef2fe16e1e99789" Dec 01 16:37:30 crc kubenswrapper[4739]: I1201 16:37:30.477801 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:37:30 crc kubenswrapper[4739]: E1201 16:37:30.478470 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:37:41 crc kubenswrapper[4739]: I1201 16:37:41.477623 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:37:41 crc kubenswrapper[4739]: E1201 16:37:41.478663 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:37:52 crc kubenswrapper[4739]: I1201 16:37:52.477949 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:37:52 crc kubenswrapper[4739]: E1201 16:37:52.478873 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:38:06 crc kubenswrapper[4739]: I1201 16:38:06.477838 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:38:06 crc kubenswrapper[4739]: E1201 16:38:06.478946 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:38:17 crc kubenswrapper[4739]: I1201 16:38:17.478147 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:38:17 crc kubenswrapper[4739]: E1201 16:38:17.479162 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:38:29 crc kubenswrapper[4739]: I1201 16:38:29.477566 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:38:29 crc kubenswrapper[4739]: E1201 16:38:29.478393 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.824329 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5khvd"] Dec 01 16:38:33 crc kubenswrapper[4739]: E1201 16:38:33.825776 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1" containerName="registry-server" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.825793 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1" containerName="registry-server" Dec 01 16:38:33 crc kubenswrapper[4739]: E1201 16:38:33.825817 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1" containerName="extract-utilities" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.825828 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1" containerName="extract-utilities" Dec 01 16:38:33 crc kubenswrapper[4739]: E1201 16:38:33.825850 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875e9c09-a5e1-455a-b12e-e46ea43be9ed" containerName="extract-content" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.825858 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="875e9c09-a5e1-455a-b12e-e46ea43be9ed" containerName="extract-content" Dec 01 16:38:33 crc kubenswrapper[4739]: E1201 16:38:33.825878 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1" containerName="extract-content" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.825887 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1" containerName="extract-content" Dec 01 16:38:33 crc kubenswrapper[4739]: E1201 16:38:33.825905 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875e9c09-a5e1-455a-b12e-e46ea43be9ed" containerName="registry-server" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.825913 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="875e9c09-a5e1-455a-b12e-e46ea43be9ed" containerName="registry-server" Dec 01 16:38:33 crc kubenswrapper[4739]: E1201 16:38:33.825939 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875e9c09-a5e1-455a-b12e-e46ea43be9ed" containerName="extract-utilities" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.825946 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="875e9c09-a5e1-455a-b12e-e46ea43be9ed" containerName="extract-utilities" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.826179 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="875e9c09-a5e1-455a-b12e-e46ea43be9ed" containerName="registry-server" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.826199 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa72de75-7cb3-4ce4-a5ee-bbac3604d8b1" containerName="registry-server" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.827828 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.853318 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5khvd"] Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.867201 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83f953ef-e08d-46b9-95bf-792107d1dfaf-utilities\") pod \"community-operators-5khvd\" (UID: \"83f953ef-e08d-46b9-95bf-792107d1dfaf\") " pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.867319 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdm7n\" (UniqueName: \"kubernetes.io/projected/83f953ef-e08d-46b9-95bf-792107d1dfaf-kube-api-access-fdm7n\") pod \"community-operators-5khvd\" (UID: \"83f953ef-e08d-46b9-95bf-792107d1dfaf\") " pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.867373 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83f953ef-e08d-46b9-95bf-792107d1dfaf-catalog-content\") pod \"community-operators-5khvd\" (UID: \"83f953ef-e08d-46b9-95bf-792107d1dfaf\") " pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.969212 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdm7n\" (UniqueName: \"kubernetes.io/projected/83f953ef-e08d-46b9-95bf-792107d1dfaf-kube-api-access-fdm7n\") pod \"community-operators-5khvd\" (UID: \"83f953ef-e08d-46b9-95bf-792107d1dfaf\") " pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.969288 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83f953ef-e08d-46b9-95bf-792107d1dfaf-catalog-content\") pod \"community-operators-5khvd\" (UID: \"83f953ef-e08d-46b9-95bf-792107d1dfaf\") " pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.969345 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83f953ef-e08d-46b9-95bf-792107d1dfaf-utilities\") pod \"community-operators-5khvd\" (UID: \"83f953ef-e08d-46b9-95bf-792107d1dfaf\") " pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.969800 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83f953ef-e08d-46b9-95bf-792107d1dfaf-utilities\") pod \"community-operators-5khvd\" (UID: \"83f953ef-e08d-46b9-95bf-792107d1dfaf\") " pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.970129 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83f953ef-e08d-46b9-95bf-792107d1dfaf-catalog-content\") pod \"community-operators-5khvd\" (UID: \"83f953ef-e08d-46b9-95bf-792107d1dfaf\") " pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:33 crc kubenswrapper[4739]: I1201 16:38:33.989122 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdm7n\" (UniqueName: \"kubernetes.io/projected/83f953ef-e08d-46b9-95bf-792107d1dfaf-kube-api-access-fdm7n\") pod \"community-operators-5khvd\" (UID: \"83f953ef-e08d-46b9-95bf-792107d1dfaf\") " pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:34 crc kubenswrapper[4739]: I1201 16:38:34.149570 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:34 crc kubenswrapper[4739]: I1201 16:38:34.693537 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5khvd"] Dec 01 16:38:35 crc kubenswrapper[4739]: I1201 16:38:35.815644 4739 generic.go:334] "Generic (PLEG): container finished" podID="83f953ef-e08d-46b9-95bf-792107d1dfaf" containerID="0ec372eddc47156357ee813f32a633512169073a32992a5f357e2b807be21eb9" exitCode=0 Dec 01 16:38:35 crc kubenswrapper[4739]: I1201 16:38:35.815907 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5khvd" event={"ID":"83f953ef-e08d-46b9-95bf-792107d1dfaf","Type":"ContainerDied","Data":"0ec372eddc47156357ee813f32a633512169073a32992a5f357e2b807be21eb9"} Dec 01 16:38:35 crc kubenswrapper[4739]: I1201 16:38:35.815931 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5khvd" event={"ID":"83f953ef-e08d-46b9-95bf-792107d1dfaf","Type":"ContainerStarted","Data":"e15e9f067193f8b130fc1371f875c59ea30cfde09bf6a8dded13318c625efc17"} Dec 01 16:38:35 crc kubenswrapper[4739]: I1201 16:38:35.817820 4739 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:38:37 crc kubenswrapper[4739]: I1201 16:38:37.834748 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5khvd" event={"ID":"83f953ef-e08d-46b9-95bf-792107d1dfaf","Type":"ContainerStarted","Data":"46279b411dacdab2d9578b36f2c352deaaf725faad2eca103797aa4fb63d5873"} Dec 01 16:38:38 crc kubenswrapper[4739]: I1201 16:38:38.844169 4739 generic.go:334] "Generic (PLEG): container finished" podID="83f953ef-e08d-46b9-95bf-792107d1dfaf" containerID="46279b411dacdab2d9578b36f2c352deaaf725faad2eca103797aa4fb63d5873" exitCode=0 Dec 01 16:38:38 crc kubenswrapper[4739]: I1201 16:38:38.844207 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5khvd" event={"ID":"83f953ef-e08d-46b9-95bf-792107d1dfaf","Type":"ContainerDied","Data":"46279b411dacdab2d9578b36f2c352deaaf725faad2eca103797aa4fb63d5873"} Dec 01 16:38:39 crc kubenswrapper[4739]: I1201 16:38:39.854642 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5khvd" event={"ID":"83f953ef-e08d-46b9-95bf-792107d1dfaf","Type":"ContainerStarted","Data":"fb93a2bfeea1241d88742a5fe3d814aeefd2b2eaf9808fdf18ddcd89d94499d7"} Dec 01 16:38:39 crc kubenswrapper[4739]: I1201 16:38:39.879037 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5khvd" podStartSLOduration=3.264234128 podStartE2EDuration="6.879014845s" podCreationTimestamp="2025-12-01 16:38:33 +0000 UTC" firstStartedPulling="2025-12-01 16:38:35.817403785 +0000 UTC m=+3817.643149879" lastFinishedPulling="2025-12-01 16:38:39.432184492 +0000 UTC m=+3821.257930596" observedRunningTime="2025-12-01 16:38:39.873882497 +0000 UTC m=+3821.699628601" watchObservedRunningTime="2025-12-01 16:38:39.879014845 +0000 UTC m=+3821.704760949" Dec 01 16:38:43 crc kubenswrapper[4739]: I1201 16:38:43.478550 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:38:43 crc kubenswrapper[4739]: E1201 16:38:43.479616 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:38:44 crc kubenswrapper[4739]: I1201 16:38:44.150647 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:44 crc kubenswrapper[4739]: I1201 16:38:44.151157 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:44 crc kubenswrapper[4739]: I1201 16:38:44.758106 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:44 crc kubenswrapper[4739]: I1201 16:38:44.959308 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:45 crc kubenswrapper[4739]: I1201 16:38:45.010935 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5khvd"] Dec 01 16:38:46 crc kubenswrapper[4739]: I1201 16:38:46.930127 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5khvd" podUID="83f953ef-e08d-46b9-95bf-792107d1dfaf" containerName="registry-server" containerID="cri-o://fb93a2bfeea1241d88742a5fe3d814aeefd2b2eaf9808fdf18ddcd89d94499d7" gracePeriod=2 Dec 01 16:38:47 crc kubenswrapper[4739]: E1201 16:38:47.064771 4739 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83f953ef_e08d_46b9_95bf_792107d1dfaf.slice/crio-fb93a2bfeea1241d88742a5fe3d814aeefd2b2eaf9808fdf18ddcd89d94499d7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83f953ef_e08d_46b9_95bf_792107d1dfaf.slice/crio-conmon-fb93a2bfeea1241d88742a5fe3d814aeefd2b2eaf9808fdf18ddcd89d94499d7.scope\": RecentStats: unable to find data in memory cache]" Dec 01 16:38:47 crc kubenswrapper[4739]: I1201 16:38:47.669865 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:47 crc kubenswrapper[4739]: I1201 16:38:47.815995 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83f953ef-e08d-46b9-95bf-792107d1dfaf-utilities\") pod \"83f953ef-e08d-46b9-95bf-792107d1dfaf\" (UID: \"83f953ef-e08d-46b9-95bf-792107d1dfaf\") " Dec 01 16:38:47 crc kubenswrapper[4739]: I1201 16:38:47.816092 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdm7n\" (UniqueName: \"kubernetes.io/projected/83f953ef-e08d-46b9-95bf-792107d1dfaf-kube-api-access-fdm7n\") pod \"83f953ef-e08d-46b9-95bf-792107d1dfaf\" (UID: \"83f953ef-e08d-46b9-95bf-792107d1dfaf\") " Dec 01 16:38:47 crc kubenswrapper[4739]: I1201 16:38:47.816184 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83f953ef-e08d-46b9-95bf-792107d1dfaf-catalog-content\") pod \"83f953ef-e08d-46b9-95bf-792107d1dfaf\" (UID: \"83f953ef-e08d-46b9-95bf-792107d1dfaf\") " Dec 01 16:38:47 crc kubenswrapper[4739]: I1201 16:38:47.817510 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83f953ef-e08d-46b9-95bf-792107d1dfaf-utilities" (OuterVolumeSpecName: "utilities") pod "83f953ef-e08d-46b9-95bf-792107d1dfaf" (UID: "83f953ef-e08d-46b9-95bf-792107d1dfaf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:38:47 crc kubenswrapper[4739]: I1201 16:38:47.893345 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83f953ef-e08d-46b9-95bf-792107d1dfaf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83f953ef-e08d-46b9-95bf-792107d1dfaf" (UID: "83f953ef-e08d-46b9-95bf-792107d1dfaf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:38:47 crc kubenswrapper[4739]: I1201 16:38:47.918322 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83f953ef-e08d-46b9-95bf-792107d1dfaf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:38:47 crc kubenswrapper[4739]: I1201 16:38:47.918363 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83f953ef-e08d-46b9-95bf-792107d1dfaf-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:38:47 crc kubenswrapper[4739]: I1201 16:38:47.938587 4739 generic.go:334] "Generic (PLEG): container finished" podID="83f953ef-e08d-46b9-95bf-792107d1dfaf" containerID="fb93a2bfeea1241d88742a5fe3d814aeefd2b2eaf9808fdf18ddcd89d94499d7" exitCode=0 Dec 01 16:38:47 crc kubenswrapper[4739]: I1201 16:38:47.938650 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5khvd" event={"ID":"83f953ef-e08d-46b9-95bf-792107d1dfaf","Type":"ContainerDied","Data":"fb93a2bfeea1241d88742a5fe3d814aeefd2b2eaf9808fdf18ddcd89d94499d7"} Dec 01 16:38:47 crc kubenswrapper[4739]: I1201 16:38:47.938673 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5khvd" Dec 01 16:38:47 crc kubenswrapper[4739]: I1201 16:38:47.938698 4739 scope.go:117] "RemoveContainer" containerID="fb93a2bfeea1241d88742a5fe3d814aeefd2b2eaf9808fdf18ddcd89d94499d7" Dec 01 16:38:47 crc kubenswrapper[4739]: I1201 16:38:47.938685 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5khvd" event={"ID":"83f953ef-e08d-46b9-95bf-792107d1dfaf","Type":"ContainerDied","Data":"e15e9f067193f8b130fc1371f875c59ea30cfde09bf6a8dded13318c625efc17"} Dec 01 16:38:47 crc kubenswrapper[4739]: I1201 16:38:47.958501 4739 scope.go:117] "RemoveContainer" containerID="46279b411dacdab2d9578b36f2c352deaaf725faad2eca103797aa4fb63d5873" Dec 01 16:38:48 crc kubenswrapper[4739]: I1201 16:38:48.474606 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83f953ef-e08d-46b9-95bf-792107d1dfaf-kube-api-access-fdm7n" (OuterVolumeSpecName: "kube-api-access-fdm7n") pod "83f953ef-e08d-46b9-95bf-792107d1dfaf" (UID: "83f953ef-e08d-46b9-95bf-792107d1dfaf"). InnerVolumeSpecName "kube-api-access-fdm7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:38:48 crc kubenswrapper[4739]: I1201 16:38:48.492733 4739 scope.go:117] "RemoveContainer" containerID="0ec372eddc47156357ee813f32a633512169073a32992a5f357e2b807be21eb9" Dec 01 16:38:48 crc kubenswrapper[4739]: I1201 16:38:48.531891 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdm7n\" (UniqueName: \"kubernetes.io/projected/83f953ef-e08d-46b9-95bf-792107d1dfaf-kube-api-access-fdm7n\") on node \"crc\" DevicePath \"\"" Dec 01 16:38:48 crc kubenswrapper[4739]: I1201 16:38:48.568162 4739 scope.go:117] "RemoveContainer" containerID="fb93a2bfeea1241d88742a5fe3d814aeefd2b2eaf9808fdf18ddcd89d94499d7" Dec 01 16:38:48 crc kubenswrapper[4739]: E1201 16:38:48.569733 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb93a2bfeea1241d88742a5fe3d814aeefd2b2eaf9808fdf18ddcd89d94499d7\": container with ID starting with fb93a2bfeea1241d88742a5fe3d814aeefd2b2eaf9808fdf18ddcd89d94499d7 not found: ID does not exist" containerID="fb93a2bfeea1241d88742a5fe3d814aeefd2b2eaf9808fdf18ddcd89d94499d7" Dec 01 16:38:48 crc kubenswrapper[4739]: I1201 16:38:48.569891 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb93a2bfeea1241d88742a5fe3d814aeefd2b2eaf9808fdf18ddcd89d94499d7"} err="failed to get container status \"fb93a2bfeea1241d88742a5fe3d814aeefd2b2eaf9808fdf18ddcd89d94499d7\": rpc error: code = NotFound desc = could not find container \"fb93a2bfeea1241d88742a5fe3d814aeefd2b2eaf9808fdf18ddcd89d94499d7\": container with ID starting with fb93a2bfeea1241d88742a5fe3d814aeefd2b2eaf9808fdf18ddcd89d94499d7 not found: ID does not exist" Dec 01 16:38:48 crc kubenswrapper[4739]: I1201 16:38:48.570025 4739 scope.go:117] "RemoveContainer" containerID="46279b411dacdab2d9578b36f2c352deaaf725faad2eca103797aa4fb63d5873" Dec 01 16:38:48 crc kubenswrapper[4739]: E1201 16:38:48.573560 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46279b411dacdab2d9578b36f2c352deaaf725faad2eca103797aa4fb63d5873\": container with ID starting with 46279b411dacdab2d9578b36f2c352deaaf725faad2eca103797aa4fb63d5873 not found: ID does not exist" containerID="46279b411dacdab2d9578b36f2c352deaaf725faad2eca103797aa4fb63d5873" Dec 01 16:38:48 crc kubenswrapper[4739]: I1201 16:38:48.573623 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46279b411dacdab2d9578b36f2c352deaaf725faad2eca103797aa4fb63d5873"} err="failed to get container status \"46279b411dacdab2d9578b36f2c352deaaf725faad2eca103797aa4fb63d5873\": rpc error: code = NotFound desc = could not find container \"46279b411dacdab2d9578b36f2c352deaaf725faad2eca103797aa4fb63d5873\": container with ID starting with 46279b411dacdab2d9578b36f2c352deaaf725faad2eca103797aa4fb63d5873 not found: ID does not exist" Dec 01 16:38:48 crc kubenswrapper[4739]: I1201 16:38:48.573664 4739 scope.go:117] "RemoveContainer" containerID="0ec372eddc47156357ee813f32a633512169073a32992a5f357e2b807be21eb9" Dec 01 16:38:48 crc kubenswrapper[4739]: E1201 16:38:48.576064 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ec372eddc47156357ee813f32a633512169073a32992a5f357e2b807be21eb9\": container with ID starting with 0ec372eddc47156357ee813f32a633512169073a32992a5f357e2b807be21eb9 not found: ID does not exist" containerID="0ec372eddc47156357ee813f32a633512169073a32992a5f357e2b807be21eb9" Dec 01 16:38:48 crc kubenswrapper[4739]: I1201 16:38:48.576164 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ec372eddc47156357ee813f32a633512169073a32992a5f357e2b807be21eb9"} err="failed to get container status \"0ec372eddc47156357ee813f32a633512169073a32992a5f357e2b807be21eb9\": rpc error: code = NotFound desc = could not find container \"0ec372eddc47156357ee813f32a633512169073a32992a5f357e2b807be21eb9\": container with ID starting with 0ec372eddc47156357ee813f32a633512169073a32992a5f357e2b807be21eb9 not found: ID does not exist" Dec 01 16:38:48 crc kubenswrapper[4739]: I1201 16:38:48.623641 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5khvd"] Dec 01 16:38:48 crc kubenswrapper[4739]: I1201 16:38:48.632274 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5khvd"] Dec 01 16:38:50 crc kubenswrapper[4739]: I1201 16:38:50.494495 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83f953ef-e08d-46b9-95bf-792107d1dfaf" path="/var/lib/kubelet/pods/83f953ef-e08d-46b9-95bf-792107d1dfaf/volumes" Dec 01 16:38:56 crc kubenswrapper[4739]: I1201 16:38:56.477500 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:38:56 crc kubenswrapper[4739]: E1201 16:38:56.478703 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:39:08 crc kubenswrapper[4739]: I1201 16:39:08.491727 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:39:08 crc kubenswrapper[4739]: E1201 16:39:08.492939 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:39:20 crc kubenswrapper[4739]: I1201 16:39:20.477249 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:39:21 crc kubenswrapper[4739]: I1201 16:39:21.283299 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"61090cedfcd49a3200345e2fe0990d802fe64985971533d0b449ee59ab888705"} Dec 01 16:41:39 crc kubenswrapper[4739]: I1201 16:41:39.621608 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:41:39 crc kubenswrapper[4739]: I1201 16:41:39.622107 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:42:09 crc kubenswrapper[4739]: I1201 16:42:09.621925 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:42:09 crc kubenswrapper[4739]: I1201 16:42:09.622668 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:42:39 crc kubenswrapper[4739]: I1201 16:42:39.621879 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:42:39 crc kubenswrapper[4739]: I1201 16:42:39.622580 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:42:39 crc kubenswrapper[4739]: I1201 16:42:39.622638 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 16:42:39 crc kubenswrapper[4739]: I1201 16:42:39.623603 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"61090cedfcd49a3200345e2fe0990d802fe64985971533d0b449ee59ab888705"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:42:39 crc kubenswrapper[4739]: I1201 16:42:39.623666 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://61090cedfcd49a3200345e2fe0990d802fe64985971533d0b449ee59ab888705" gracePeriod=600 Dec 01 16:42:40 crc kubenswrapper[4739]: I1201 16:42:40.316583 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="61090cedfcd49a3200345e2fe0990d802fe64985971533d0b449ee59ab888705" exitCode=0 Dec 01 16:42:40 crc kubenswrapper[4739]: I1201 16:42:40.317164 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"61090cedfcd49a3200345e2fe0990d802fe64985971533d0b449ee59ab888705"} Dec 01 16:42:40 crc kubenswrapper[4739]: I1201 16:42:40.317193 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59"} Dec 01 16:42:40 crc kubenswrapper[4739]: I1201 16:42:40.317210 4739 scope.go:117] "RemoveContainer" containerID="ca709c5d2413bd01679009a80410b3ca3abbc3ac2b4074ee2b5a78ebdd1ae364" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.513291 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ld7zz"] Dec 01 16:43:17 crc kubenswrapper[4739]: E1201 16:43:17.514166 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f953ef-e08d-46b9-95bf-792107d1dfaf" containerName="extract-utilities" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.514178 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f953ef-e08d-46b9-95bf-792107d1dfaf" containerName="extract-utilities" Dec 01 16:43:17 crc kubenswrapper[4739]: E1201 16:43:17.514194 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f953ef-e08d-46b9-95bf-792107d1dfaf" containerName="extract-content" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.514200 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f953ef-e08d-46b9-95bf-792107d1dfaf" containerName="extract-content" Dec 01 16:43:17 crc kubenswrapper[4739]: E1201 16:43:17.514228 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f953ef-e08d-46b9-95bf-792107d1dfaf" containerName="registry-server" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.514236 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f953ef-e08d-46b9-95bf-792107d1dfaf" containerName="registry-server" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.514409 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f953ef-e08d-46b9-95bf-792107d1dfaf" containerName="registry-server" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.515731 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.522082 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ld7zz"] Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.555881 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqgsn\" (UniqueName: \"kubernetes.io/projected/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-kube-api-access-fqgsn\") pod \"redhat-operators-ld7zz\" (UID: \"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81\") " pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.555954 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-catalog-content\") pod \"redhat-operators-ld7zz\" (UID: \"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81\") " pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.556058 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-utilities\") pod \"redhat-operators-ld7zz\" (UID: \"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81\") " pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.658271 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-utilities\") pod \"redhat-operators-ld7zz\" (UID: \"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81\") " pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.658353 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-utilities\") pod \"redhat-operators-ld7zz\" (UID: \"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81\") " pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.658548 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqgsn\" (UniqueName: \"kubernetes.io/projected/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-kube-api-access-fqgsn\") pod \"redhat-operators-ld7zz\" (UID: \"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81\") " pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.658598 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-catalog-content\") pod \"redhat-operators-ld7zz\" (UID: \"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81\") " pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.658959 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-catalog-content\") pod \"redhat-operators-ld7zz\" (UID: \"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81\") " pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.681258 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqgsn\" (UniqueName: \"kubernetes.io/projected/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-kube-api-access-fqgsn\") pod \"redhat-operators-ld7zz\" (UID: \"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81\") " pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:17 crc kubenswrapper[4739]: I1201 16:43:17.841871 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:18 crc kubenswrapper[4739]: I1201 16:43:18.328357 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ld7zz"] Dec 01 16:43:18 crc kubenswrapper[4739]: I1201 16:43:18.753524 4739 generic.go:334] "Generic (PLEG): container finished" podID="4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" containerID="7bcb8b8a706b3574cc051ebf5a36808c05a768985b6c4f3f3805e248182146c0" exitCode=0 Dec 01 16:43:18 crc kubenswrapper[4739]: I1201 16:43:18.753817 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ld7zz" event={"ID":"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81","Type":"ContainerDied","Data":"7bcb8b8a706b3574cc051ebf5a36808c05a768985b6c4f3f3805e248182146c0"} Dec 01 16:43:18 crc kubenswrapper[4739]: I1201 16:43:18.753843 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ld7zz" event={"ID":"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81","Type":"ContainerStarted","Data":"3bfbcdbc9e654b04b82bb26220abcd764076456176315673420324af8fbd04cc"} Dec 01 16:43:19 crc kubenswrapper[4739]: I1201 16:43:19.762659 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ld7zz" event={"ID":"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81","Type":"ContainerStarted","Data":"a9c9ada0dd9a1244fcaa5ec9d7419cec79e2c2e983b94e0e1a751723ca23942f"} Dec 01 16:43:21 crc kubenswrapper[4739]: I1201 16:43:21.786473 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ld7zz" event={"ID":"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81","Type":"ContainerDied","Data":"a9c9ada0dd9a1244fcaa5ec9d7419cec79e2c2e983b94e0e1a751723ca23942f"} Dec 01 16:43:21 crc kubenswrapper[4739]: I1201 16:43:21.788195 4739 generic.go:334] "Generic (PLEG): container finished" podID="4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" containerID="a9c9ada0dd9a1244fcaa5ec9d7419cec79e2c2e983b94e0e1a751723ca23942f" exitCode=0 Dec 01 16:43:22 crc kubenswrapper[4739]: I1201 16:43:22.799504 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ld7zz" event={"ID":"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81","Type":"ContainerStarted","Data":"8bfa1189a7afc2579e3f187e4e4632f151defc9cbd8689e6093a3a03f9b4ebd4"} Dec 01 16:43:22 crc kubenswrapper[4739]: I1201 16:43:22.826431 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ld7zz" podStartSLOduration=2.332238562 podStartE2EDuration="5.8263896s" podCreationTimestamp="2025-12-01 16:43:17 +0000 UTC" firstStartedPulling="2025-12-01 16:43:18.755217878 +0000 UTC m=+4100.580963972" lastFinishedPulling="2025-12-01 16:43:22.249368876 +0000 UTC m=+4104.075115010" observedRunningTime="2025-12-01 16:43:22.817392312 +0000 UTC m=+4104.643138426" watchObservedRunningTime="2025-12-01 16:43:22.8263896 +0000 UTC m=+4104.652135704" Dec 01 16:43:27 crc kubenswrapper[4739]: I1201 16:43:27.842175 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:27 crc kubenswrapper[4739]: I1201 16:43:27.842798 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:28 crc kubenswrapper[4739]: I1201 16:43:28.888951 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ld7zz" podUID="4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" containerName="registry-server" probeResult="failure" output=< Dec 01 16:43:28 crc kubenswrapper[4739]: timeout: failed to connect service ":50051" within 1s Dec 01 16:43:28 crc kubenswrapper[4739]: > Dec 01 16:43:37 crc kubenswrapper[4739]: I1201 16:43:37.904128 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:37 crc kubenswrapper[4739]: I1201 16:43:37.967560 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:38 crc kubenswrapper[4739]: I1201 16:43:38.144345 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ld7zz"] Dec 01 16:43:38 crc kubenswrapper[4739]: I1201 16:43:38.960251 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ld7zz" podUID="4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" containerName="registry-server" containerID="cri-o://8bfa1189a7afc2579e3f187e4e4632f151defc9cbd8689e6093a3a03f9b4ebd4" gracePeriod=2 Dec 01 16:43:39 crc kubenswrapper[4739]: I1201 16:43:39.970643 4739 generic.go:334] "Generic (PLEG): container finished" podID="4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" containerID="8bfa1189a7afc2579e3f187e4e4632f151defc9cbd8689e6093a3a03f9b4ebd4" exitCode=0 Dec 01 16:43:39 crc kubenswrapper[4739]: I1201 16:43:39.970724 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ld7zz" event={"ID":"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81","Type":"ContainerDied","Data":"8bfa1189a7afc2579e3f187e4e4632f151defc9cbd8689e6093a3a03f9b4ebd4"} Dec 01 16:43:40 crc kubenswrapper[4739]: I1201 16:43:40.133836 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:40 crc kubenswrapper[4739]: I1201 16:43:40.189950 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqgsn\" (UniqueName: \"kubernetes.io/projected/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-kube-api-access-fqgsn\") pod \"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81\" (UID: \"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81\") " Dec 01 16:43:40 crc kubenswrapper[4739]: I1201 16:43:40.190050 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-utilities\") pod \"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81\" (UID: \"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81\") " Dec 01 16:43:40 crc kubenswrapper[4739]: I1201 16:43:40.190102 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-catalog-content\") pod \"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81\" (UID: \"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81\") " Dec 01 16:43:40 crc kubenswrapper[4739]: I1201 16:43:40.191196 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-utilities" (OuterVolumeSpecName: "utilities") pod "4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" (UID: "4ad3d6be-b42f-4431-afb1-c35c6b4e3b81"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:43:40 crc kubenswrapper[4739]: I1201 16:43:40.197908 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-kube-api-access-fqgsn" (OuterVolumeSpecName: "kube-api-access-fqgsn") pod "4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" (UID: "4ad3d6be-b42f-4431-afb1-c35c6b4e3b81"). InnerVolumeSpecName "kube-api-access-fqgsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:43:40 crc kubenswrapper[4739]: I1201 16:43:40.276340 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" (UID: "4ad3d6be-b42f-4431-afb1-c35c6b4e3b81"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:43:40 crc kubenswrapper[4739]: I1201 16:43:40.292790 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:43:40 crc kubenswrapper[4739]: I1201 16:43:40.292819 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqgsn\" (UniqueName: \"kubernetes.io/projected/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-kube-api-access-fqgsn\") on node \"crc\" DevicePath \"\"" Dec 01 16:43:40 crc kubenswrapper[4739]: I1201 16:43:40.292847 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:43:40 crc kubenswrapper[4739]: I1201 16:43:40.985174 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ld7zz" event={"ID":"4ad3d6be-b42f-4431-afb1-c35c6b4e3b81","Type":"ContainerDied","Data":"3bfbcdbc9e654b04b82bb26220abcd764076456176315673420324af8fbd04cc"} Dec 01 16:43:40 crc kubenswrapper[4739]: I1201 16:43:40.985585 4739 scope.go:117] "RemoveContainer" containerID="8bfa1189a7afc2579e3f187e4e4632f151defc9cbd8689e6093a3a03f9b4ebd4" Dec 01 16:43:40 crc kubenswrapper[4739]: I1201 16:43:40.985270 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ld7zz" Dec 01 16:43:41 crc kubenswrapper[4739]: I1201 16:43:41.010294 4739 scope.go:117] "RemoveContainer" containerID="a9c9ada0dd9a1244fcaa5ec9d7419cec79e2c2e983b94e0e1a751723ca23942f" Dec 01 16:43:41 crc kubenswrapper[4739]: I1201 16:43:41.032833 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ld7zz"] Dec 01 16:43:41 crc kubenswrapper[4739]: I1201 16:43:41.040556 4739 scope.go:117] "RemoveContainer" containerID="7bcb8b8a706b3574cc051ebf5a36808c05a768985b6c4f3f3805e248182146c0" Dec 01 16:43:41 crc kubenswrapper[4739]: I1201 16:43:41.053632 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ld7zz"] Dec 01 16:43:42 crc kubenswrapper[4739]: I1201 16:43:42.494883 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" path="/var/lib/kubelet/pods/4ad3d6be-b42f-4431-afb1-c35c6b4e3b81/volumes" Dec 01 16:43:56 crc kubenswrapper[4739]: I1201 16:43:56.821527 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bwvrm"] Dec 01 16:43:56 crc kubenswrapper[4739]: E1201 16:43:56.824199 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" containerName="registry-server" Dec 01 16:43:56 crc kubenswrapper[4739]: I1201 16:43:56.824370 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" containerName="registry-server" Dec 01 16:43:56 crc kubenswrapper[4739]: E1201 16:43:56.824572 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" containerName="extract-utilities" Dec 01 16:43:56 crc kubenswrapper[4739]: I1201 16:43:56.824728 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" containerName="extract-utilities" Dec 01 16:43:56 crc kubenswrapper[4739]: E1201 16:43:56.825813 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" containerName="extract-content" Dec 01 16:43:56 crc kubenswrapper[4739]: I1201 16:43:56.826009 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" containerName="extract-content" Dec 01 16:43:56 crc kubenswrapper[4739]: I1201 16:43:56.827713 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ad3d6be-b42f-4431-afb1-c35c6b4e3b81" containerName="registry-server" Dec 01 16:43:56 crc kubenswrapper[4739]: I1201 16:43:56.842706 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:43:56 crc kubenswrapper[4739]: I1201 16:43:56.860721 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bwvrm"] Dec 01 16:43:56 crc kubenswrapper[4739]: I1201 16:43:56.995187 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-catalog-content\") pod \"certified-operators-bwvrm\" (UID: \"1c5fe35c-45d2-426e-9bdb-3957ad900d5b\") " pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:43:56 crc kubenswrapper[4739]: I1201 16:43:56.995356 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhzr9\" (UniqueName: \"kubernetes.io/projected/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-kube-api-access-lhzr9\") pod \"certified-operators-bwvrm\" (UID: \"1c5fe35c-45d2-426e-9bdb-3957ad900d5b\") " pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:43:56 crc kubenswrapper[4739]: I1201 16:43:56.995574 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-utilities\") pod \"certified-operators-bwvrm\" (UID: \"1c5fe35c-45d2-426e-9bdb-3957ad900d5b\") " pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:43:57 crc kubenswrapper[4739]: I1201 16:43:57.098002 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-catalog-content\") pod \"certified-operators-bwvrm\" (UID: \"1c5fe35c-45d2-426e-9bdb-3957ad900d5b\") " pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:43:57 crc kubenswrapper[4739]: I1201 16:43:57.098103 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhzr9\" (UniqueName: \"kubernetes.io/projected/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-kube-api-access-lhzr9\") pod \"certified-operators-bwvrm\" (UID: \"1c5fe35c-45d2-426e-9bdb-3957ad900d5b\") " pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:43:57 crc kubenswrapper[4739]: I1201 16:43:57.098227 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-utilities\") pod \"certified-operators-bwvrm\" (UID: \"1c5fe35c-45d2-426e-9bdb-3957ad900d5b\") " pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:43:57 crc kubenswrapper[4739]: I1201 16:43:57.098612 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-catalog-content\") pod \"certified-operators-bwvrm\" (UID: \"1c5fe35c-45d2-426e-9bdb-3957ad900d5b\") " pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:43:57 crc kubenswrapper[4739]: I1201 16:43:57.098669 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-utilities\") pod \"certified-operators-bwvrm\" (UID: \"1c5fe35c-45d2-426e-9bdb-3957ad900d5b\") " pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:43:57 crc kubenswrapper[4739]: I1201 16:43:57.119011 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhzr9\" (UniqueName: \"kubernetes.io/projected/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-kube-api-access-lhzr9\") pod \"certified-operators-bwvrm\" (UID: \"1c5fe35c-45d2-426e-9bdb-3957ad900d5b\") " pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:43:57 crc kubenswrapper[4739]: I1201 16:43:57.181646 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:43:57 crc kubenswrapper[4739]: I1201 16:43:57.710222 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bwvrm"] Dec 01 16:43:58 crc kubenswrapper[4739]: I1201 16:43:58.175871 4739 generic.go:334] "Generic (PLEG): container finished" podID="1c5fe35c-45d2-426e-9bdb-3957ad900d5b" containerID="4a302c55d1237dafe8eb8ddef0055236e52c84ddff79765ed2b95b03ed951618" exitCode=0 Dec 01 16:43:58 crc kubenswrapper[4739]: I1201 16:43:58.175931 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bwvrm" event={"ID":"1c5fe35c-45d2-426e-9bdb-3957ad900d5b","Type":"ContainerDied","Data":"4a302c55d1237dafe8eb8ddef0055236e52c84ddff79765ed2b95b03ed951618"} Dec 01 16:43:58 crc kubenswrapper[4739]: I1201 16:43:58.176180 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bwvrm" event={"ID":"1c5fe35c-45d2-426e-9bdb-3957ad900d5b","Type":"ContainerStarted","Data":"80c27f80e6729b489fb3aeeaa0c1550ac960c6c06fac0bb112e6212229900155"} Dec 01 16:43:58 crc kubenswrapper[4739]: I1201 16:43:58.179818 4739 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:43:59 crc kubenswrapper[4739]: I1201 16:43:59.198443 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bwvrm" event={"ID":"1c5fe35c-45d2-426e-9bdb-3957ad900d5b","Type":"ContainerStarted","Data":"c74a005dd73d42a3a7ec8101c18424e8f593fd370a83a04e417f1e8f1ff55b73"} Dec 01 16:43:59 crc kubenswrapper[4739]: I1201 16:43:59.614660 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fkd86"] Dec 01 16:43:59 crc kubenswrapper[4739]: I1201 16:43:59.622890 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:43:59 crc kubenswrapper[4739]: I1201 16:43:59.651174 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fkd86"] Dec 01 16:43:59 crc kubenswrapper[4739]: I1201 16:43:59.749990 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a513dc-5ddc-4df0-ad21-59127e9bd630-catalog-content\") pod \"redhat-marketplace-fkd86\" (UID: \"a4a513dc-5ddc-4df0-ad21-59127e9bd630\") " pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:43:59 crc kubenswrapper[4739]: I1201 16:43:59.750367 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqg2r\" (UniqueName: \"kubernetes.io/projected/a4a513dc-5ddc-4df0-ad21-59127e9bd630-kube-api-access-rqg2r\") pod \"redhat-marketplace-fkd86\" (UID: \"a4a513dc-5ddc-4df0-ad21-59127e9bd630\") " pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:43:59 crc kubenswrapper[4739]: I1201 16:43:59.750437 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a513dc-5ddc-4df0-ad21-59127e9bd630-utilities\") pod \"redhat-marketplace-fkd86\" (UID: \"a4a513dc-5ddc-4df0-ad21-59127e9bd630\") " pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:43:59 crc kubenswrapper[4739]: I1201 16:43:59.851855 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a513dc-5ddc-4df0-ad21-59127e9bd630-catalog-content\") pod \"redhat-marketplace-fkd86\" (UID: \"a4a513dc-5ddc-4df0-ad21-59127e9bd630\") " pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:43:59 crc kubenswrapper[4739]: I1201 16:43:59.851907 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqg2r\" (UniqueName: \"kubernetes.io/projected/a4a513dc-5ddc-4df0-ad21-59127e9bd630-kube-api-access-rqg2r\") pod \"redhat-marketplace-fkd86\" (UID: \"a4a513dc-5ddc-4df0-ad21-59127e9bd630\") " pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:43:59 crc kubenswrapper[4739]: I1201 16:43:59.851940 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a513dc-5ddc-4df0-ad21-59127e9bd630-utilities\") pod \"redhat-marketplace-fkd86\" (UID: \"a4a513dc-5ddc-4df0-ad21-59127e9bd630\") " pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:43:59 crc kubenswrapper[4739]: I1201 16:43:59.852460 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a513dc-5ddc-4df0-ad21-59127e9bd630-utilities\") pod \"redhat-marketplace-fkd86\" (UID: \"a4a513dc-5ddc-4df0-ad21-59127e9bd630\") " pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:43:59 crc kubenswrapper[4739]: I1201 16:43:59.852460 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a513dc-5ddc-4df0-ad21-59127e9bd630-catalog-content\") pod \"redhat-marketplace-fkd86\" (UID: \"a4a513dc-5ddc-4df0-ad21-59127e9bd630\") " pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:43:59 crc kubenswrapper[4739]: I1201 16:43:59.877645 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqg2r\" (UniqueName: \"kubernetes.io/projected/a4a513dc-5ddc-4df0-ad21-59127e9bd630-kube-api-access-rqg2r\") pod \"redhat-marketplace-fkd86\" (UID: \"a4a513dc-5ddc-4df0-ad21-59127e9bd630\") " pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:43:59 crc kubenswrapper[4739]: I1201 16:43:59.956077 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:44:00 crc kubenswrapper[4739]: I1201 16:44:00.209262 4739 generic.go:334] "Generic (PLEG): container finished" podID="1c5fe35c-45d2-426e-9bdb-3957ad900d5b" containerID="c74a005dd73d42a3a7ec8101c18424e8f593fd370a83a04e417f1e8f1ff55b73" exitCode=0 Dec 01 16:44:00 crc kubenswrapper[4739]: I1201 16:44:00.209550 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bwvrm" event={"ID":"1c5fe35c-45d2-426e-9bdb-3957ad900d5b","Type":"ContainerDied","Data":"c74a005dd73d42a3a7ec8101c18424e8f593fd370a83a04e417f1e8f1ff55b73"} Dec 01 16:44:00 crc kubenswrapper[4739]: I1201 16:44:00.425993 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fkd86"] Dec 01 16:44:01 crc kubenswrapper[4739]: I1201 16:44:01.227633 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bwvrm" event={"ID":"1c5fe35c-45d2-426e-9bdb-3957ad900d5b","Type":"ContainerStarted","Data":"6759bb28fc2c905c61e55797e0489c168c67f4ab4c0a19ec0767082eaff87f94"} Dec 01 16:44:01 crc kubenswrapper[4739]: I1201 16:44:01.229335 4739 generic.go:334] "Generic (PLEG): container finished" podID="a4a513dc-5ddc-4df0-ad21-59127e9bd630" containerID="29f6ed3bea8d4e1c493dc5b0a0f7c3d01f4678adff08fa91ee5e78fb3ecbc303" exitCode=0 Dec 01 16:44:01 crc kubenswrapper[4739]: I1201 16:44:01.229401 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fkd86" event={"ID":"a4a513dc-5ddc-4df0-ad21-59127e9bd630","Type":"ContainerDied","Data":"29f6ed3bea8d4e1c493dc5b0a0f7c3d01f4678adff08fa91ee5e78fb3ecbc303"} Dec 01 16:44:01 crc kubenswrapper[4739]: I1201 16:44:01.229522 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fkd86" event={"ID":"a4a513dc-5ddc-4df0-ad21-59127e9bd630","Type":"ContainerStarted","Data":"057f954d6ea7fcdc279d591bcaa159ec301db6da74c1fb4260413e5279bc20d2"} Dec 01 16:44:01 crc kubenswrapper[4739]: I1201 16:44:01.252161 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bwvrm" podStartSLOduration=2.675179221 podStartE2EDuration="5.252148035s" podCreationTimestamp="2025-12-01 16:43:56 +0000 UTC" firstStartedPulling="2025-12-01 16:43:58.179372475 +0000 UTC m=+4140.005118579" lastFinishedPulling="2025-12-01 16:44:00.756341289 +0000 UTC m=+4142.582087393" observedRunningTime="2025-12-01 16:44:01.249907376 +0000 UTC m=+4143.075653510" watchObservedRunningTime="2025-12-01 16:44:01.252148035 +0000 UTC m=+4143.077894129" Dec 01 16:44:03 crc kubenswrapper[4739]: I1201 16:44:03.260561 4739 generic.go:334] "Generic (PLEG): container finished" podID="a4a513dc-5ddc-4df0-ad21-59127e9bd630" containerID="04a18dffde05be416954cad7bd810f56a0770f74ae03bef8b39ff89905413ff0" exitCode=0 Dec 01 16:44:03 crc kubenswrapper[4739]: I1201 16:44:03.261060 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fkd86" event={"ID":"a4a513dc-5ddc-4df0-ad21-59127e9bd630","Type":"ContainerDied","Data":"04a18dffde05be416954cad7bd810f56a0770f74ae03bef8b39ff89905413ff0"} Dec 01 16:44:04 crc kubenswrapper[4739]: I1201 16:44:04.271081 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fkd86" event={"ID":"a4a513dc-5ddc-4df0-ad21-59127e9bd630","Type":"ContainerStarted","Data":"dc55a9990e9bae8c227e8a34ce517449586049f77edea9d082264c2e648c2071"} Dec 01 16:44:04 crc kubenswrapper[4739]: I1201 16:44:04.299032 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fkd86" podStartSLOduration=2.81945269 podStartE2EDuration="5.299005095s" podCreationTimestamp="2025-12-01 16:43:59 +0000 UTC" firstStartedPulling="2025-12-01 16:44:01.230747754 +0000 UTC m=+4143.056493848" lastFinishedPulling="2025-12-01 16:44:03.710300099 +0000 UTC m=+4145.536046253" observedRunningTime="2025-12-01 16:44:04.293079402 +0000 UTC m=+4146.118825496" watchObservedRunningTime="2025-12-01 16:44:04.299005095 +0000 UTC m=+4146.124751229" Dec 01 16:44:07 crc kubenswrapper[4739]: I1201 16:44:07.182910 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:44:07 crc kubenswrapper[4739]: I1201 16:44:07.183666 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:44:07 crc kubenswrapper[4739]: I1201 16:44:07.273276 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:44:07 crc kubenswrapper[4739]: I1201 16:44:07.348778 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:44:07 crc kubenswrapper[4739]: I1201 16:44:07.600691 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bwvrm"] Dec 01 16:44:09 crc kubenswrapper[4739]: I1201 16:44:09.317975 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bwvrm" podUID="1c5fe35c-45d2-426e-9bdb-3957ad900d5b" containerName="registry-server" containerID="cri-o://6759bb28fc2c905c61e55797e0489c168c67f4ab4c0a19ec0767082eaff87f94" gracePeriod=2 Dec 01 16:44:09 crc kubenswrapper[4739]: I1201 16:44:09.949941 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:44:09 crc kubenswrapper[4739]: I1201 16:44:09.956244 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:44:09 crc kubenswrapper[4739]: I1201 16:44:09.956294 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.014556 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.077914 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhzr9\" (UniqueName: \"kubernetes.io/projected/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-kube-api-access-lhzr9\") pod \"1c5fe35c-45d2-426e-9bdb-3957ad900d5b\" (UID: \"1c5fe35c-45d2-426e-9bdb-3957ad900d5b\") " Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.078061 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-catalog-content\") pod \"1c5fe35c-45d2-426e-9bdb-3957ad900d5b\" (UID: \"1c5fe35c-45d2-426e-9bdb-3957ad900d5b\") " Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.078209 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-utilities\") pod \"1c5fe35c-45d2-426e-9bdb-3957ad900d5b\" (UID: \"1c5fe35c-45d2-426e-9bdb-3957ad900d5b\") " Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.084401 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-utilities" (OuterVolumeSpecName: "utilities") pod "1c5fe35c-45d2-426e-9bdb-3957ad900d5b" (UID: "1c5fe35c-45d2-426e-9bdb-3957ad900d5b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.084639 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-kube-api-access-lhzr9" (OuterVolumeSpecName: "kube-api-access-lhzr9") pod "1c5fe35c-45d2-426e-9bdb-3957ad900d5b" (UID: "1c5fe35c-45d2-426e-9bdb-3957ad900d5b"). InnerVolumeSpecName "kube-api-access-lhzr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.182373 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.182884 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhzr9\" (UniqueName: \"kubernetes.io/projected/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-kube-api-access-lhzr9\") on node \"crc\" DevicePath \"\"" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.193407 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c5fe35c-45d2-426e-9bdb-3957ad900d5b" (UID: "1c5fe35c-45d2-426e-9bdb-3957ad900d5b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.284710 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c5fe35c-45d2-426e-9bdb-3957ad900d5b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.327131 4739 generic.go:334] "Generic (PLEG): container finished" podID="1c5fe35c-45d2-426e-9bdb-3957ad900d5b" containerID="6759bb28fc2c905c61e55797e0489c168c67f4ab4c0a19ec0767082eaff87f94" exitCode=0 Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.327245 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bwvrm" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.328147 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bwvrm" event={"ID":"1c5fe35c-45d2-426e-9bdb-3957ad900d5b","Type":"ContainerDied","Data":"6759bb28fc2c905c61e55797e0489c168c67f4ab4c0a19ec0767082eaff87f94"} Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.328267 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bwvrm" event={"ID":"1c5fe35c-45d2-426e-9bdb-3957ad900d5b","Type":"ContainerDied","Data":"80c27f80e6729b489fb3aeeaa0c1550ac960c6c06fac0bb112e6212229900155"} Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.328391 4739 scope.go:117] "RemoveContainer" containerID="6759bb28fc2c905c61e55797e0489c168c67f4ab4c0a19ec0767082eaff87f94" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.354490 4739 scope.go:117] "RemoveContainer" containerID="c74a005dd73d42a3a7ec8101c18424e8f593fd370a83a04e417f1e8f1ff55b73" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.368378 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bwvrm"] Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.375669 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bwvrm"] Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.390063 4739 scope.go:117] "RemoveContainer" containerID="4a302c55d1237dafe8eb8ddef0055236e52c84ddff79765ed2b95b03ed951618" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.406904 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.449111 4739 scope.go:117] "RemoveContainer" containerID="6759bb28fc2c905c61e55797e0489c168c67f4ab4c0a19ec0767082eaff87f94" Dec 01 16:44:10 crc kubenswrapper[4739]: E1201 16:44:10.450613 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6759bb28fc2c905c61e55797e0489c168c67f4ab4c0a19ec0767082eaff87f94\": container with ID starting with 6759bb28fc2c905c61e55797e0489c168c67f4ab4c0a19ec0767082eaff87f94 not found: ID does not exist" containerID="6759bb28fc2c905c61e55797e0489c168c67f4ab4c0a19ec0767082eaff87f94" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.450664 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6759bb28fc2c905c61e55797e0489c168c67f4ab4c0a19ec0767082eaff87f94"} err="failed to get container status \"6759bb28fc2c905c61e55797e0489c168c67f4ab4c0a19ec0767082eaff87f94\": rpc error: code = NotFound desc = could not find container \"6759bb28fc2c905c61e55797e0489c168c67f4ab4c0a19ec0767082eaff87f94\": container with ID starting with 6759bb28fc2c905c61e55797e0489c168c67f4ab4c0a19ec0767082eaff87f94 not found: ID does not exist" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.450696 4739 scope.go:117] "RemoveContainer" containerID="c74a005dd73d42a3a7ec8101c18424e8f593fd370a83a04e417f1e8f1ff55b73" Dec 01 16:44:10 crc kubenswrapper[4739]: E1201 16:44:10.451779 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c74a005dd73d42a3a7ec8101c18424e8f593fd370a83a04e417f1e8f1ff55b73\": container with ID starting with c74a005dd73d42a3a7ec8101c18424e8f593fd370a83a04e417f1e8f1ff55b73 not found: ID does not exist" containerID="c74a005dd73d42a3a7ec8101c18424e8f593fd370a83a04e417f1e8f1ff55b73" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.451830 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c74a005dd73d42a3a7ec8101c18424e8f593fd370a83a04e417f1e8f1ff55b73"} err="failed to get container status \"c74a005dd73d42a3a7ec8101c18424e8f593fd370a83a04e417f1e8f1ff55b73\": rpc error: code = NotFound desc = could not find container \"c74a005dd73d42a3a7ec8101c18424e8f593fd370a83a04e417f1e8f1ff55b73\": container with ID starting with c74a005dd73d42a3a7ec8101c18424e8f593fd370a83a04e417f1e8f1ff55b73 not found: ID does not exist" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.451863 4739 scope.go:117] "RemoveContainer" containerID="4a302c55d1237dafe8eb8ddef0055236e52c84ddff79765ed2b95b03ed951618" Dec 01 16:44:10 crc kubenswrapper[4739]: E1201 16:44:10.455134 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a302c55d1237dafe8eb8ddef0055236e52c84ddff79765ed2b95b03ed951618\": container with ID starting with 4a302c55d1237dafe8eb8ddef0055236e52c84ddff79765ed2b95b03ed951618 not found: ID does not exist" containerID="4a302c55d1237dafe8eb8ddef0055236e52c84ddff79765ed2b95b03ed951618" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.455973 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a302c55d1237dafe8eb8ddef0055236e52c84ddff79765ed2b95b03ed951618"} err="failed to get container status \"4a302c55d1237dafe8eb8ddef0055236e52c84ddff79765ed2b95b03ed951618\": rpc error: code = NotFound desc = could not find container \"4a302c55d1237dafe8eb8ddef0055236e52c84ddff79765ed2b95b03ed951618\": container with ID starting with 4a302c55d1237dafe8eb8ddef0055236e52c84ddff79765ed2b95b03ed951618 not found: ID does not exist" Dec 01 16:44:10 crc kubenswrapper[4739]: I1201 16:44:10.498767 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c5fe35c-45d2-426e-9bdb-3957ad900d5b" path="/var/lib/kubelet/pods/1c5fe35c-45d2-426e-9bdb-3957ad900d5b/volumes" Dec 01 16:44:12 crc kubenswrapper[4739]: I1201 16:44:12.408033 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fkd86"] Dec 01 16:44:12 crc kubenswrapper[4739]: I1201 16:44:12.408756 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fkd86" podUID="a4a513dc-5ddc-4df0-ad21-59127e9bd630" containerName="registry-server" containerID="cri-o://dc55a9990e9bae8c227e8a34ce517449586049f77edea9d082264c2e648c2071" gracePeriod=2 Dec 01 16:44:12 crc kubenswrapper[4739]: I1201 16:44:12.942038 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.050311 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a513dc-5ddc-4df0-ad21-59127e9bd630-catalog-content\") pod \"a4a513dc-5ddc-4df0-ad21-59127e9bd630\" (UID: \"a4a513dc-5ddc-4df0-ad21-59127e9bd630\") " Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.050377 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a513dc-5ddc-4df0-ad21-59127e9bd630-utilities\") pod \"a4a513dc-5ddc-4df0-ad21-59127e9bd630\" (UID: \"a4a513dc-5ddc-4df0-ad21-59127e9bd630\") " Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.050442 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqg2r\" (UniqueName: \"kubernetes.io/projected/a4a513dc-5ddc-4df0-ad21-59127e9bd630-kube-api-access-rqg2r\") pod \"a4a513dc-5ddc-4df0-ad21-59127e9bd630\" (UID: \"a4a513dc-5ddc-4df0-ad21-59127e9bd630\") " Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.051504 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4a513dc-5ddc-4df0-ad21-59127e9bd630-utilities" (OuterVolumeSpecName: "utilities") pod "a4a513dc-5ddc-4df0-ad21-59127e9bd630" (UID: "a4a513dc-5ddc-4df0-ad21-59127e9bd630"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.063679 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4a513dc-5ddc-4df0-ad21-59127e9bd630-kube-api-access-rqg2r" (OuterVolumeSpecName: "kube-api-access-rqg2r") pod "a4a513dc-5ddc-4df0-ad21-59127e9bd630" (UID: "a4a513dc-5ddc-4df0-ad21-59127e9bd630"). InnerVolumeSpecName "kube-api-access-rqg2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.078963 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4a513dc-5ddc-4df0-ad21-59127e9bd630-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4a513dc-5ddc-4df0-ad21-59127e9bd630" (UID: "a4a513dc-5ddc-4df0-ad21-59127e9bd630"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.152555 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a513dc-5ddc-4df0-ad21-59127e9bd630-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.152592 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a513dc-5ddc-4df0-ad21-59127e9bd630-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.152602 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqg2r\" (UniqueName: \"kubernetes.io/projected/a4a513dc-5ddc-4df0-ad21-59127e9bd630-kube-api-access-rqg2r\") on node \"crc\" DevicePath \"\"" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.364606 4739 generic.go:334] "Generic (PLEG): container finished" podID="a4a513dc-5ddc-4df0-ad21-59127e9bd630" containerID="dc55a9990e9bae8c227e8a34ce517449586049f77edea9d082264c2e648c2071" exitCode=0 Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.364654 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fkd86" event={"ID":"a4a513dc-5ddc-4df0-ad21-59127e9bd630","Type":"ContainerDied","Data":"dc55a9990e9bae8c227e8a34ce517449586049f77edea9d082264c2e648c2071"} Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.364682 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fkd86" event={"ID":"a4a513dc-5ddc-4df0-ad21-59127e9bd630","Type":"ContainerDied","Data":"057f954d6ea7fcdc279d591bcaa159ec301db6da74c1fb4260413e5279bc20d2"} Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.364703 4739 scope.go:117] "RemoveContainer" containerID="dc55a9990e9bae8c227e8a34ce517449586049f77edea9d082264c2e648c2071" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.367570 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fkd86" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.390457 4739 scope.go:117] "RemoveContainer" containerID="04a18dffde05be416954cad7bd810f56a0770f74ae03bef8b39ff89905413ff0" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.423512 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fkd86"] Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.438000 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fkd86"] Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.445806 4739 scope.go:117] "RemoveContainer" containerID="29f6ed3bea8d4e1c493dc5b0a0f7c3d01f4678adff08fa91ee5e78fb3ecbc303" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.486001 4739 scope.go:117] "RemoveContainer" containerID="dc55a9990e9bae8c227e8a34ce517449586049f77edea9d082264c2e648c2071" Dec 01 16:44:13 crc kubenswrapper[4739]: E1201 16:44:13.486369 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc55a9990e9bae8c227e8a34ce517449586049f77edea9d082264c2e648c2071\": container with ID starting with dc55a9990e9bae8c227e8a34ce517449586049f77edea9d082264c2e648c2071 not found: ID does not exist" containerID="dc55a9990e9bae8c227e8a34ce517449586049f77edea9d082264c2e648c2071" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.486407 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc55a9990e9bae8c227e8a34ce517449586049f77edea9d082264c2e648c2071"} err="failed to get container status \"dc55a9990e9bae8c227e8a34ce517449586049f77edea9d082264c2e648c2071\": rpc error: code = NotFound desc = could not find container \"dc55a9990e9bae8c227e8a34ce517449586049f77edea9d082264c2e648c2071\": container with ID starting with dc55a9990e9bae8c227e8a34ce517449586049f77edea9d082264c2e648c2071 not found: ID does not exist" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.486448 4739 scope.go:117] "RemoveContainer" containerID="04a18dffde05be416954cad7bd810f56a0770f74ae03bef8b39ff89905413ff0" Dec 01 16:44:13 crc kubenswrapper[4739]: E1201 16:44:13.486707 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04a18dffde05be416954cad7bd810f56a0770f74ae03bef8b39ff89905413ff0\": container with ID starting with 04a18dffde05be416954cad7bd810f56a0770f74ae03bef8b39ff89905413ff0 not found: ID does not exist" containerID="04a18dffde05be416954cad7bd810f56a0770f74ae03bef8b39ff89905413ff0" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.486784 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04a18dffde05be416954cad7bd810f56a0770f74ae03bef8b39ff89905413ff0"} err="failed to get container status \"04a18dffde05be416954cad7bd810f56a0770f74ae03bef8b39ff89905413ff0\": rpc error: code = NotFound desc = could not find container \"04a18dffde05be416954cad7bd810f56a0770f74ae03bef8b39ff89905413ff0\": container with ID starting with 04a18dffde05be416954cad7bd810f56a0770f74ae03bef8b39ff89905413ff0 not found: ID does not exist" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.486847 4739 scope.go:117] "RemoveContainer" containerID="29f6ed3bea8d4e1c493dc5b0a0f7c3d01f4678adff08fa91ee5e78fb3ecbc303" Dec 01 16:44:13 crc kubenswrapper[4739]: E1201 16:44:13.487287 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29f6ed3bea8d4e1c493dc5b0a0f7c3d01f4678adff08fa91ee5e78fb3ecbc303\": container with ID starting with 29f6ed3bea8d4e1c493dc5b0a0f7c3d01f4678adff08fa91ee5e78fb3ecbc303 not found: ID does not exist" containerID="29f6ed3bea8d4e1c493dc5b0a0f7c3d01f4678adff08fa91ee5e78fb3ecbc303" Dec 01 16:44:13 crc kubenswrapper[4739]: I1201 16:44:13.487312 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29f6ed3bea8d4e1c493dc5b0a0f7c3d01f4678adff08fa91ee5e78fb3ecbc303"} err="failed to get container status \"29f6ed3bea8d4e1c493dc5b0a0f7c3d01f4678adff08fa91ee5e78fb3ecbc303\": rpc error: code = NotFound desc = could not find container \"29f6ed3bea8d4e1c493dc5b0a0f7c3d01f4678adff08fa91ee5e78fb3ecbc303\": container with ID starting with 29f6ed3bea8d4e1c493dc5b0a0f7c3d01f4678adff08fa91ee5e78fb3ecbc303 not found: ID does not exist" Dec 01 16:44:14 crc kubenswrapper[4739]: I1201 16:44:14.495534 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4a513dc-5ddc-4df0-ad21-59127e9bd630" path="/var/lib/kubelet/pods/a4a513dc-5ddc-4df0-ad21-59127e9bd630/volumes" Dec 01 16:44:39 crc kubenswrapper[4739]: I1201 16:44:39.621914 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:44:39 crc kubenswrapper[4739]: I1201 16:44:39.622950 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.206099 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj"] Dec 01 16:45:00 crc kubenswrapper[4739]: E1201 16:45:00.208724 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c5fe35c-45d2-426e-9bdb-3957ad900d5b" containerName="registry-server" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.208740 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c5fe35c-45d2-426e-9bdb-3957ad900d5b" containerName="registry-server" Dec 01 16:45:00 crc kubenswrapper[4739]: E1201 16:45:00.208762 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4a513dc-5ddc-4df0-ad21-59127e9bd630" containerName="extract-utilities" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.208768 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4a513dc-5ddc-4df0-ad21-59127e9bd630" containerName="extract-utilities" Dec 01 16:45:00 crc kubenswrapper[4739]: E1201 16:45:00.208779 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4a513dc-5ddc-4df0-ad21-59127e9bd630" containerName="extract-content" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.208785 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4a513dc-5ddc-4df0-ad21-59127e9bd630" containerName="extract-content" Dec 01 16:45:00 crc kubenswrapper[4739]: E1201 16:45:00.208809 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4a513dc-5ddc-4df0-ad21-59127e9bd630" containerName="registry-server" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.208814 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4a513dc-5ddc-4df0-ad21-59127e9bd630" containerName="registry-server" Dec 01 16:45:00 crc kubenswrapper[4739]: E1201 16:45:00.208825 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c5fe35c-45d2-426e-9bdb-3957ad900d5b" containerName="extract-content" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.208830 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c5fe35c-45d2-426e-9bdb-3957ad900d5b" containerName="extract-content" Dec 01 16:45:00 crc kubenswrapper[4739]: E1201 16:45:00.208846 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c5fe35c-45d2-426e-9bdb-3957ad900d5b" containerName="extract-utilities" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.208852 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c5fe35c-45d2-426e-9bdb-3957ad900d5b" containerName="extract-utilities" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.209007 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4a513dc-5ddc-4df0-ad21-59127e9bd630" containerName="registry-server" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.209028 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c5fe35c-45d2-426e-9bdb-3957ad900d5b" containerName="registry-server" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.209673 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.217187 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.217954 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.222991 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj"] Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.383743 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab3e9a22-f81f-43a5-be51-873db28f6fc5-secret-volume\") pod \"collect-profiles-29410125-tdzmj\" (UID: \"ab3e9a22-f81f-43a5-be51-873db28f6fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.383891 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab3e9a22-f81f-43a5-be51-873db28f6fc5-config-volume\") pod \"collect-profiles-29410125-tdzmj\" (UID: \"ab3e9a22-f81f-43a5-be51-873db28f6fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.384099 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj5lg\" (UniqueName: \"kubernetes.io/projected/ab3e9a22-f81f-43a5-be51-873db28f6fc5-kube-api-access-kj5lg\") pod \"collect-profiles-29410125-tdzmj\" (UID: \"ab3e9a22-f81f-43a5-be51-873db28f6fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.485589 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab3e9a22-f81f-43a5-be51-873db28f6fc5-secret-volume\") pod \"collect-profiles-29410125-tdzmj\" (UID: \"ab3e9a22-f81f-43a5-be51-873db28f6fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.486047 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab3e9a22-f81f-43a5-be51-873db28f6fc5-config-volume\") pod \"collect-profiles-29410125-tdzmj\" (UID: \"ab3e9a22-f81f-43a5-be51-873db28f6fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.486181 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj5lg\" (UniqueName: \"kubernetes.io/projected/ab3e9a22-f81f-43a5-be51-873db28f6fc5-kube-api-access-kj5lg\") pod \"collect-profiles-29410125-tdzmj\" (UID: \"ab3e9a22-f81f-43a5-be51-873db28f6fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.487431 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab3e9a22-f81f-43a5-be51-873db28f6fc5-config-volume\") pod \"collect-profiles-29410125-tdzmj\" (UID: \"ab3e9a22-f81f-43a5-be51-873db28f6fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.513309 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab3e9a22-f81f-43a5-be51-873db28f6fc5-secret-volume\") pod \"collect-profiles-29410125-tdzmj\" (UID: \"ab3e9a22-f81f-43a5-be51-873db28f6fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.513310 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj5lg\" (UniqueName: \"kubernetes.io/projected/ab3e9a22-f81f-43a5-be51-873db28f6fc5-kube-api-access-kj5lg\") pod \"collect-profiles-29410125-tdzmj\" (UID: \"ab3e9a22-f81f-43a5-be51-873db28f6fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" Dec 01 16:45:00 crc kubenswrapper[4739]: I1201 16:45:00.531984 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" Dec 01 16:45:01 crc kubenswrapper[4739]: I1201 16:45:01.021265 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj"] Dec 01 16:45:01 crc kubenswrapper[4739]: I1201 16:45:01.214026 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" event={"ID":"ab3e9a22-f81f-43a5-be51-873db28f6fc5","Type":"ContainerStarted","Data":"7c42de16425bc019ee7aec24bfd23900b48082518e5b1a6ceadd90a6a4851604"} Dec 01 16:45:02 crc kubenswrapper[4739]: I1201 16:45:02.232013 4739 generic.go:334] "Generic (PLEG): container finished" podID="ab3e9a22-f81f-43a5-be51-873db28f6fc5" containerID="d9e460a75e7b7a687aa69972602f0dac295c8227aa50757cc3bbe36bacff0f92" exitCode=0 Dec 01 16:45:02 crc kubenswrapper[4739]: I1201 16:45:02.232102 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" event={"ID":"ab3e9a22-f81f-43a5-be51-873db28f6fc5","Type":"ContainerDied","Data":"d9e460a75e7b7a687aa69972602f0dac295c8227aa50757cc3bbe36bacff0f92"} Dec 01 16:45:04 crc kubenswrapper[4739]: I1201 16:45:04.587926 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" Dec 01 16:45:04 crc kubenswrapper[4739]: I1201 16:45:04.687819 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab3e9a22-f81f-43a5-be51-873db28f6fc5-config-volume\") pod \"ab3e9a22-f81f-43a5-be51-873db28f6fc5\" (UID: \"ab3e9a22-f81f-43a5-be51-873db28f6fc5\") " Dec 01 16:45:04 crc kubenswrapper[4739]: I1201 16:45:04.687906 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab3e9a22-f81f-43a5-be51-873db28f6fc5-secret-volume\") pod \"ab3e9a22-f81f-43a5-be51-873db28f6fc5\" (UID: \"ab3e9a22-f81f-43a5-be51-873db28f6fc5\") " Dec 01 16:45:04 crc kubenswrapper[4739]: I1201 16:45:04.687983 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kj5lg\" (UniqueName: \"kubernetes.io/projected/ab3e9a22-f81f-43a5-be51-873db28f6fc5-kube-api-access-kj5lg\") pod \"ab3e9a22-f81f-43a5-be51-873db28f6fc5\" (UID: \"ab3e9a22-f81f-43a5-be51-873db28f6fc5\") " Dec 01 16:45:04 crc kubenswrapper[4739]: I1201 16:45:04.689444 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab3e9a22-f81f-43a5-be51-873db28f6fc5-config-volume" (OuterVolumeSpecName: "config-volume") pod "ab3e9a22-f81f-43a5-be51-873db28f6fc5" (UID: "ab3e9a22-f81f-43a5-be51-873db28f6fc5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:45:04 crc kubenswrapper[4739]: I1201 16:45:04.700215 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab3e9a22-f81f-43a5-be51-873db28f6fc5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ab3e9a22-f81f-43a5-be51-873db28f6fc5" (UID: "ab3e9a22-f81f-43a5-be51-873db28f6fc5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:45:04 crc kubenswrapper[4739]: I1201 16:45:04.711466 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab3e9a22-f81f-43a5-be51-873db28f6fc5-kube-api-access-kj5lg" (OuterVolumeSpecName: "kube-api-access-kj5lg") pod "ab3e9a22-f81f-43a5-be51-873db28f6fc5" (UID: "ab3e9a22-f81f-43a5-be51-873db28f6fc5"). InnerVolumeSpecName "kube-api-access-kj5lg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:45:04 crc kubenswrapper[4739]: I1201 16:45:04.790076 4739 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ab3e9a22-f81f-43a5-be51-873db28f6fc5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:45:04 crc kubenswrapper[4739]: I1201 16:45:04.790108 4739 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ab3e9a22-f81f-43a5-be51-873db28f6fc5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:45:04 crc kubenswrapper[4739]: I1201 16:45:04.790117 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kj5lg\" (UniqueName: \"kubernetes.io/projected/ab3e9a22-f81f-43a5-be51-873db28f6fc5-kube-api-access-kj5lg\") on node \"crc\" DevicePath \"\"" Dec 01 16:45:05 crc kubenswrapper[4739]: I1201 16:45:05.265464 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" event={"ID":"ab3e9a22-f81f-43a5-be51-873db28f6fc5","Type":"ContainerDied","Data":"7c42de16425bc019ee7aec24bfd23900b48082518e5b1a6ceadd90a6a4851604"} Dec 01 16:45:05 crc kubenswrapper[4739]: I1201 16:45:05.265516 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c42de16425bc019ee7aec24bfd23900b48082518e5b1a6ceadd90a6a4851604" Dec 01 16:45:05 crc kubenswrapper[4739]: I1201 16:45:05.265547 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-tdzmj" Dec 01 16:45:05 crc kubenswrapper[4739]: I1201 16:45:05.686453 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq"] Dec 01 16:45:05 crc kubenswrapper[4739]: I1201 16:45:05.699740 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410080-7tttq"] Dec 01 16:45:06 crc kubenswrapper[4739]: I1201 16:45:06.501068 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74b76c6e-6b96-4361-b887-8b3d3779235c" path="/var/lib/kubelet/pods/74b76c6e-6b96-4361-b887-8b3d3779235c/volumes" Dec 01 16:45:09 crc kubenswrapper[4739]: I1201 16:45:09.622242 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:45:09 crc kubenswrapper[4739]: I1201 16:45:09.622790 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:45:25 crc kubenswrapper[4739]: I1201 16:45:25.168067 4739 scope.go:117] "RemoveContainer" containerID="033a74d0d0712e84a4423c97a7a17ee50087258363dec676c9a8a2f68c4cb433" Dec 01 16:45:39 crc kubenswrapper[4739]: I1201 16:45:39.621657 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:45:39 crc kubenswrapper[4739]: I1201 16:45:39.622339 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:45:39 crc kubenswrapper[4739]: I1201 16:45:39.622405 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 16:45:39 crc kubenswrapper[4739]: I1201 16:45:39.623676 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:45:39 crc kubenswrapper[4739]: I1201 16:45:39.623793 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" gracePeriod=600 Dec 01 16:45:40 crc kubenswrapper[4739]: E1201 16:45:40.271096 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:45:40 crc kubenswrapper[4739]: I1201 16:45:40.623617 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" exitCode=0 Dec 01 16:45:40 crc kubenswrapper[4739]: I1201 16:45:40.623682 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59"} Dec 01 16:45:40 crc kubenswrapper[4739]: I1201 16:45:40.623731 4739 scope.go:117] "RemoveContainer" containerID="61090cedfcd49a3200345e2fe0990d802fe64985971533d0b449ee59ab888705" Dec 01 16:45:40 crc kubenswrapper[4739]: I1201 16:45:40.624628 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:45:40 crc kubenswrapper[4739]: E1201 16:45:40.625142 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:45:55 crc kubenswrapper[4739]: I1201 16:45:55.477731 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:45:55 crc kubenswrapper[4739]: E1201 16:45:55.478837 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:46:06 crc kubenswrapper[4739]: I1201 16:46:06.479186 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:46:06 crc kubenswrapper[4739]: E1201 16:46:06.480070 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:46:21 crc kubenswrapper[4739]: I1201 16:46:21.477727 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:46:21 crc kubenswrapper[4739]: E1201 16:46:21.479774 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:46:32 crc kubenswrapper[4739]: I1201 16:46:32.477880 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:46:32 crc kubenswrapper[4739]: E1201 16:46:32.478796 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:46:47 crc kubenswrapper[4739]: I1201 16:46:47.477110 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:46:47 crc kubenswrapper[4739]: E1201 16:46:47.478208 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:47:01 crc kubenswrapper[4739]: I1201 16:47:01.477330 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:47:01 crc kubenswrapper[4739]: E1201 16:47:01.478074 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:47:13 crc kubenswrapper[4739]: I1201 16:47:13.478361 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:47:13 crc kubenswrapper[4739]: E1201 16:47:13.479519 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:47:26 crc kubenswrapper[4739]: I1201 16:47:26.478056 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:47:26 crc kubenswrapper[4739]: E1201 16:47:26.479389 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:47:37 crc kubenswrapper[4739]: I1201 16:47:37.478233 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:47:37 crc kubenswrapper[4739]: E1201 16:47:37.479251 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:47:49 crc kubenswrapper[4739]: I1201 16:47:49.477333 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:47:49 crc kubenswrapper[4739]: E1201 16:47:49.478049 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:48:04 crc kubenswrapper[4739]: I1201 16:48:04.477376 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:48:04 crc kubenswrapper[4739]: E1201 16:48:04.479129 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:48:17 crc kubenswrapper[4739]: I1201 16:48:17.478106 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:48:17 crc kubenswrapper[4739]: E1201 16:48:17.479213 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:48:28 crc kubenswrapper[4739]: I1201 16:48:28.488886 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:48:28 crc kubenswrapper[4739]: E1201 16:48:28.489893 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:48:43 crc kubenswrapper[4739]: I1201 16:48:43.480599 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:48:43 crc kubenswrapper[4739]: E1201 16:48:43.484357 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:48:54 crc kubenswrapper[4739]: I1201 16:48:54.478876 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:48:54 crc kubenswrapper[4739]: E1201 16:48:54.480035 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:49:07 crc kubenswrapper[4739]: I1201 16:49:07.478309 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:49:07 crc kubenswrapper[4739]: E1201 16:49:07.479507 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:49:20 crc kubenswrapper[4739]: I1201 16:49:20.477468 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:49:20 crc kubenswrapper[4739]: E1201 16:49:20.478379 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:49:31 crc kubenswrapper[4739]: I1201 16:49:31.479026 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:49:31 crc kubenswrapper[4739]: E1201 16:49:31.479825 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:49:40 crc kubenswrapper[4739]: I1201 16:49:40.609338 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p7zps"] Dec 01 16:49:40 crc kubenswrapper[4739]: E1201 16:49:40.611395 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab3e9a22-f81f-43a5-be51-873db28f6fc5" containerName="collect-profiles" Dec 01 16:49:40 crc kubenswrapper[4739]: I1201 16:49:40.611546 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab3e9a22-f81f-43a5-be51-873db28f6fc5" containerName="collect-profiles" Dec 01 16:49:40 crc kubenswrapper[4739]: I1201 16:49:40.611881 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab3e9a22-f81f-43a5-be51-873db28f6fc5" containerName="collect-profiles" Dec 01 16:49:40 crc kubenswrapper[4739]: I1201 16:49:40.613611 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:49:40 crc kubenswrapper[4739]: I1201 16:49:40.677831 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p7zps"] Dec 01 16:49:40 crc kubenswrapper[4739]: I1201 16:49:40.697349 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55dad234-160d-42ec-b1b1-0d21d262c1e8-utilities\") pod \"community-operators-p7zps\" (UID: \"55dad234-160d-42ec-b1b1-0d21d262c1e8\") " pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:49:40 crc kubenswrapper[4739]: I1201 16:49:40.697756 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55dad234-160d-42ec-b1b1-0d21d262c1e8-catalog-content\") pod \"community-operators-p7zps\" (UID: \"55dad234-160d-42ec-b1b1-0d21d262c1e8\") " pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:49:40 crc kubenswrapper[4739]: I1201 16:49:40.697879 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s2lb\" (UniqueName: \"kubernetes.io/projected/55dad234-160d-42ec-b1b1-0d21d262c1e8-kube-api-access-5s2lb\") pod \"community-operators-p7zps\" (UID: \"55dad234-160d-42ec-b1b1-0d21d262c1e8\") " pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:49:40 crc kubenswrapper[4739]: I1201 16:49:40.799828 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55dad234-160d-42ec-b1b1-0d21d262c1e8-utilities\") pod \"community-operators-p7zps\" (UID: \"55dad234-160d-42ec-b1b1-0d21d262c1e8\") " pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:49:40 crc kubenswrapper[4739]: I1201 16:49:40.799991 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55dad234-160d-42ec-b1b1-0d21d262c1e8-catalog-content\") pod \"community-operators-p7zps\" (UID: \"55dad234-160d-42ec-b1b1-0d21d262c1e8\") " pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:49:40 crc kubenswrapper[4739]: I1201 16:49:40.800210 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s2lb\" (UniqueName: \"kubernetes.io/projected/55dad234-160d-42ec-b1b1-0d21d262c1e8-kube-api-access-5s2lb\") pod \"community-operators-p7zps\" (UID: \"55dad234-160d-42ec-b1b1-0d21d262c1e8\") " pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:49:40 crc kubenswrapper[4739]: I1201 16:49:40.801101 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55dad234-160d-42ec-b1b1-0d21d262c1e8-utilities\") pod \"community-operators-p7zps\" (UID: \"55dad234-160d-42ec-b1b1-0d21d262c1e8\") " pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:49:40 crc kubenswrapper[4739]: I1201 16:49:40.801381 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55dad234-160d-42ec-b1b1-0d21d262c1e8-catalog-content\") pod \"community-operators-p7zps\" (UID: \"55dad234-160d-42ec-b1b1-0d21d262c1e8\") " pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:49:40 crc kubenswrapper[4739]: I1201 16:49:40.823222 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s2lb\" (UniqueName: \"kubernetes.io/projected/55dad234-160d-42ec-b1b1-0d21d262c1e8-kube-api-access-5s2lb\") pod \"community-operators-p7zps\" (UID: \"55dad234-160d-42ec-b1b1-0d21d262c1e8\") " pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:49:40 crc kubenswrapper[4739]: I1201 16:49:40.935493 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:49:41 crc kubenswrapper[4739]: I1201 16:49:41.467683 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p7zps"] Dec 01 16:49:42 crc kubenswrapper[4739]: I1201 16:49:42.178564 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7zps" event={"ID":"55dad234-160d-42ec-b1b1-0d21d262c1e8","Type":"ContainerStarted","Data":"0faa439a27e1f82abb96161b28f187bbfbc0ba8113817c67e394f5b43a067c08"} Dec 01 16:49:42 crc kubenswrapper[4739]: I1201 16:49:42.178946 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7zps" event={"ID":"55dad234-160d-42ec-b1b1-0d21d262c1e8","Type":"ContainerStarted","Data":"881e86839082796e248d918861a459d680034f7dc2b1452183c9419df050d601"} Dec 01 16:49:43 crc kubenswrapper[4739]: I1201 16:49:43.190446 4739 generic.go:334] "Generic (PLEG): container finished" podID="55dad234-160d-42ec-b1b1-0d21d262c1e8" containerID="0faa439a27e1f82abb96161b28f187bbfbc0ba8113817c67e394f5b43a067c08" exitCode=0 Dec 01 16:49:43 crc kubenswrapper[4739]: I1201 16:49:43.190550 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7zps" event={"ID":"55dad234-160d-42ec-b1b1-0d21d262c1e8","Type":"ContainerDied","Data":"0faa439a27e1f82abb96161b28f187bbfbc0ba8113817c67e394f5b43a067c08"} Dec 01 16:49:43 crc kubenswrapper[4739]: I1201 16:49:43.193471 4739 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:49:45 crc kubenswrapper[4739]: I1201 16:49:45.226940 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7zps" event={"ID":"55dad234-160d-42ec-b1b1-0d21d262c1e8","Type":"ContainerStarted","Data":"f66dd5f20755756a882dc18a75e8b7ad4cf063b5e072f4e782a116de799c6945"} Dec 01 16:49:46 crc kubenswrapper[4739]: I1201 16:49:46.239697 4739 generic.go:334] "Generic (PLEG): container finished" podID="55dad234-160d-42ec-b1b1-0d21d262c1e8" containerID="f66dd5f20755756a882dc18a75e8b7ad4cf063b5e072f4e782a116de799c6945" exitCode=0 Dec 01 16:49:46 crc kubenswrapper[4739]: I1201 16:49:46.239995 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7zps" event={"ID":"55dad234-160d-42ec-b1b1-0d21d262c1e8","Type":"ContainerDied","Data":"f66dd5f20755756a882dc18a75e8b7ad4cf063b5e072f4e782a116de799c6945"} Dec 01 16:49:46 crc kubenswrapper[4739]: I1201 16:49:46.479328 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:49:46 crc kubenswrapper[4739]: E1201 16:49:46.480635 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:49:47 crc kubenswrapper[4739]: I1201 16:49:47.254670 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7zps" event={"ID":"55dad234-160d-42ec-b1b1-0d21d262c1e8","Type":"ContainerStarted","Data":"72809328d94f96932d80e90d2bf216d553947ef8f60387c3746285eb9261d75e"} Dec 01 16:49:47 crc kubenswrapper[4739]: I1201 16:49:47.287161 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p7zps" podStartSLOduration=3.580733922 podStartE2EDuration="7.287142299s" podCreationTimestamp="2025-12-01 16:49:40 +0000 UTC" firstStartedPulling="2025-12-01 16:49:43.193108879 +0000 UTC m=+4485.018854983" lastFinishedPulling="2025-12-01 16:49:46.899517266 +0000 UTC m=+4488.725263360" observedRunningTime="2025-12-01 16:49:47.285345524 +0000 UTC m=+4489.111091618" watchObservedRunningTime="2025-12-01 16:49:47.287142299 +0000 UTC m=+4489.112888393" Dec 01 16:49:50 crc kubenswrapper[4739]: I1201 16:49:50.936449 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:49:50 crc kubenswrapper[4739]: I1201 16:49:50.937034 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:49:50 crc kubenswrapper[4739]: I1201 16:49:50.990203 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:50:00 crc kubenswrapper[4739]: I1201 16:50:00.477723 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:50:00 crc kubenswrapper[4739]: E1201 16:50:00.478596 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:50:00 crc kubenswrapper[4739]: I1201 16:50:00.991633 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:50:01 crc kubenswrapper[4739]: I1201 16:50:01.040999 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p7zps"] Dec 01 16:50:01 crc kubenswrapper[4739]: I1201 16:50:01.371740 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p7zps" podUID="55dad234-160d-42ec-b1b1-0d21d262c1e8" containerName="registry-server" containerID="cri-o://72809328d94f96932d80e90d2bf216d553947ef8f60387c3746285eb9261d75e" gracePeriod=2 Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.071018 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.170435 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s2lb\" (UniqueName: \"kubernetes.io/projected/55dad234-160d-42ec-b1b1-0d21d262c1e8-kube-api-access-5s2lb\") pod \"55dad234-160d-42ec-b1b1-0d21d262c1e8\" (UID: \"55dad234-160d-42ec-b1b1-0d21d262c1e8\") " Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.171260 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55dad234-160d-42ec-b1b1-0d21d262c1e8-catalog-content\") pod \"55dad234-160d-42ec-b1b1-0d21d262c1e8\" (UID: \"55dad234-160d-42ec-b1b1-0d21d262c1e8\") " Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.171493 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55dad234-160d-42ec-b1b1-0d21d262c1e8-utilities\") pod \"55dad234-160d-42ec-b1b1-0d21d262c1e8\" (UID: \"55dad234-160d-42ec-b1b1-0d21d262c1e8\") " Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.172447 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55dad234-160d-42ec-b1b1-0d21d262c1e8-utilities" (OuterVolumeSpecName: "utilities") pod "55dad234-160d-42ec-b1b1-0d21d262c1e8" (UID: "55dad234-160d-42ec-b1b1-0d21d262c1e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.178246 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55dad234-160d-42ec-b1b1-0d21d262c1e8-kube-api-access-5s2lb" (OuterVolumeSpecName: "kube-api-access-5s2lb") pod "55dad234-160d-42ec-b1b1-0d21d262c1e8" (UID: "55dad234-160d-42ec-b1b1-0d21d262c1e8"). InnerVolumeSpecName "kube-api-access-5s2lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.230276 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55dad234-160d-42ec-b1b1-0d21d262c1e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55dad234-160d-42ec-b1b1-0d21d262c1e8" (UID: "55dad234-160d-42ec-b1b1-0d21d262c1e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.274034 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55dad234-160d-42ec-b1b1-0d21d262c1e8-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.274264 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s2lb\" (UniqueName: \"kubernetes.io/projected/55dad234-160d-42ec-b1b1-0d21d262c1e8-kube-api-access-5s2lb\") on node \"crc\" DevicePath \"\"" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.274334 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55dad234-160d-42ec-b1b1-0d21d262c1e8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.380594 4739 generic.go:334] "Generic (PLEG): container finished" podID="55dad234-160d-42ec-b1b1-0d21d262c1e8" containerID="72809328d94f96932d80e90d2bf216d553947ef8f60387c3746285eb9261d75e" exitCode=0 Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.380637 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7zps" event={"ID":"55dad234-160d-42ec-b1b1-0d21d262c1e8","Type":"ContainerDied","Data":"72809328d94f96932d80e90d2bf216d553947ef8f60387c3746285eb9261d75e"} Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.380664 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7zps" event={"ID":"55dad234-160d-42ec-b1b1-0d21d262c1e8","Type":"ContainerDied","Data":"881e86839082796e248d918861a459d680034f7dc2b1452183c9419df050d601"} Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.380684 4739 scope.go:117] "RemoveContainer" containerID="72809328d94f96932d80e90d2bf216d553947ef8f60387c3746285eb9261d75e" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.380823 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p7zps" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.415538 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p7zps"] Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.417887 4739 scope.go:117] "RemoveContainer" containerID="f66dd5f20755756a882dc18a75e8b7ad4cf063b5e072f4e782a116de799c6945" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.424565 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p7zps"] Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.444255 4739 scope.go:117] "RemoveContainer" containerID="0faa439a27e1f82abb96161b28f187bbfbc0ba8113817c67e394f5b43a067c08" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.484632 4739 scope.go:117] "RemoveContainer" containerID="72809328d94f96932d80e90d2bf216d553947ef8f60387c3746285eb9261d75e" Dec 01 16:50:02 crc kubenswrapper[4739]: E1201 16:50:02.484957 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72809328d94f96932d80e90d2bf216d553947ef8f60387c3746285eb9261d75e\": container with ID starting with 72809328d94f96932d80e90d2bf216d553947ef8f60387c3746285eb9261d75e not found: ID does not exist" containerID="72809328d94f96932d80e90d2bf216d553947ef8f60387c3746285eb9261d75e" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.484990 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72809328d94f96932d80e90d2bf216d553947ef8f60387c3746285eb9261d75e"} err="failed to get container status \"72809328d94f96932d80e90d2bf216d553947ef8f60387c3746285eb9261d75e\": rpc error: code = NotFound desc = could not find container \"72809328d94f96932d80e90d2bf216d553947ef8f60387c3746285eb9261d75e\": container with ID starting with 72809328d94f96932d80e90d2bf216d553947ef8f60387c3746285eb9261d75e not found: ID does not exist" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.485021 4739 scope.go:117] "RemoveContainer" containerID="f66dd5f20755756a882dc18a75e8b7ad4cf063b5e072f4e782a116de799c6945" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.487322 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55dad234-160d-42ec-b1b1-0d21d262c1e8" path="/var/lib/kubelet/pods/55dad234-160d-42ec-b1b1-0d21d262c1e8/volumes" Dec 01 16:50:02 crc kubenswrapper[4739]: E1201 16:50:02.487542 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f66dd5f20755756a882dc18a75e8b7ad4cf063b5e072f4e782a116de799c6945\": container with ID starting with f66dd5f20755756a882dc18a75e8b7ad4cf063b5e072f4e782a116de799c6945 not found: ID does not exist" containerID="f66dd5f20755756a882dc18a75e8b7ad4cf063b5e072f4e782a116de799c6945" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.487583 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f66dd5f20755756a882dc18a75e8b7ad4cf063b5e072f4e782a116de799c6945"} err="failed to get container status \"f66dd5f20755756a882dc18a75e8b7ad4cf063b5e072f4e782a116de799c6945\": rpc error: code = NotFound desc = could not find container \"f66dd5f20755756a882dc18a75e8b7ad4cf063b5e072f4e782a116de799c6945\": container with ID starting with f66dd5f20755756a882dc18a75e8b7ad4cf063b5e072f4e782a116de799c6945 not found: ID does not exist" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.487608 4739 scope.go:117] "RemoveContainer" containerID="0faa439a27e1f82abb96161b28f187bbfbc0ba8113817c67e394f5b43a067c08" Dec 01 16:50:02 crc kubenswrapper[4739]: E1201 16:50:02.488147 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0faa439a27e1f82abb96161b28f187bbfbc0ba8113817c67e394f5b43a067c08\": container with ID starting with 0faa439a27e1f82abb96161b28f187bbfbc0ba8113817c67e394f5b43a067c08 not found: ID does not exist" containerID="0faa439a27e1f82abb96161b28f187bbfbc0ba8113817c67e394f5b43a067c08" Dec 01 16:50:02 crc kubenswrapper[4739]: I1201 16:50:02.488181 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0faa439a27e1f82abb96161b28f187bbfbc0ba8113817c67e394f5b43a067c08"} err="failed to get container status \"0faa439a27e1f82abb96161b28f187bbfbc0ba8113817c67e394f5b43a067c08\": rpc error: code = NotFound desc = could not find container \"0faa439a27e1f82abb96161b28f187bbfbc0ba8113817c67e394f5b43a067c08\": container with ID starting with 0faa439a27e1f82abb96161b28f187bbfbc0ba8113817c67e394f5b43a067c08 not found: ID does not exist" Dec 01 16:50:14 crc kubenswrapper[4739]: I1201 16:50:14.477178 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:50:14 crc kubenswrapper[4739]: E1201 16:50:14.478001 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:50:29 crc kubenswrapper[4739]: I1201 16:50:29.476897 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:50:29 crc kubenswrapper[4739]: E1201 16:50:29.477548 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:50:41 crc kubenswrapper[4739]: I1201 16:50:41.478136 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:50:42 crc kubenswrapper[4739]: I1201 16:50:42.728252 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"af6de8887e1f92dd65adf332c6767d5ff238c9babbbeb1db75ee37509b77d751"} Dec 01 16:51:35 crc kubenswrapper[4739]: I1201 16:51:35.250357 4739 generic.go:334] "Generic (PLEG): container finished" podID="d7438aa0-1e45-4112-9af4-a584825d29df" containerID="cf2aad93d9175d6043838d61f159d79af1a951f700e12bddc32e32a631e77e5f" exitCode=1 Dec 01 16:51:35 crc kubenswrapper[4739]: I1201 16:51:35.250448 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d7438aa0-1e45-4112-9af4-a584825d29df","Type":"ContainerDied","Data":"cf2aad93d9175d6043838d61f159d79af1a951f700e12bddc32e32a631e77e5f"} Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.085431 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.235580 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"d7438aa0-1e45-4112-9af4-a584825d29df\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.235648 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d7438aa0-1e45-4112-9af4-a584825d29df-test-operator-ephemeral-temporary\") pod \"d7438aa0-1e45-4112-9af4-a584825d29df\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.235711 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d7438aa0-1e45-4112-9af4-a584825d29df-test-operator-ephemeral-workdir\") pod \"d7438aa0-1e45-4112-9af4-a584825d29df\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.235796 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-openstack-config-secret\") pod \"d7438aa0-1e45-4112-9af4-a584825d29df\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.235816 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-ssh-key\") pod \"d7438aa0-1e45-4112-9af4-a584825d29df\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.235838 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7438aa0-1e45-4112-9af4-a584825d29df-config-data\") pod \"d7438aa0-1e45-4112-9af4-a584825d29df\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.235891 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d7438aa0-1e45-4112-9af4-a584825d29df-openstack-config\") pod \"d7438aa0-1e45-4112-9af4-a584825d29df\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.236195 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7438aa0-1e45-4112-9af4-a584825d29df-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "d7438aa0-1e45-4112-9af4-a584825d29df" (UID: "d7438aa0-1e45-4112-9af4-a584825d29df"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.235943 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-ca-certs\") pod \"d7438aa0-1e45-4112-9af4-a584825d29df\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.236713 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvzjb\" (UniqueName: \"kubernetes.io/projected/d7438aa0-1e45-4112-9af4-a584825d29df-kube-api-access-cvzjb\") pod \"d7438aa0-1e45-4112-9af4-a584825d29df\" (UID: \"d7438aa0-1e45-4112-9af4-a584825d29df\") " Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.236791 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7438aa0-1e45-4112-9af4-a584825d29df-config-data" (OuterVolumeSpecName: "config-data") pod "d7438aa0-1e45-4112-9af4-a584825d29df" (UID: "d7438aa0-1e45-4112-9af4-a584825d29df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.237270 4739 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d7438aa0-1e45-4112-9af4-a584825d29df-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.237294 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7438aa0-1e45-4112-9af4-a584825d29df-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.241906 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7438aa0-1e45-4112-9af4-a584825d29df-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "d7438aa0-1e45-4112-9af4-a584825d29df" (UID: "d7438aa0-1e45-4112-9af4-a584825d29df"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.252207 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "d7438aa0-1e45-4112-9af4-a584825d29df" (UID: "d7438aa0-1e45-4112-9af4-a584825d29df"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.255148 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7438aa0-1e45-4112-9af4-a584825d29df-kube-api-access-cvzjb" (OuterVolumeSpecName: "kube-api-access-cvzjb") pod "d7438aa0-1e45-4112-9af4-a584825d29df" (UID: "d7438aa0-1e45-4112-9af4-a584825d29df"). InnerVolumeSpecName "kube-api-access-cvzjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.275015 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d7438aa0-1e45-4112-9af4-a584825d29df" (UID: "d7438aa0-1e45-4112-9af4-a584825d29df"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.280833 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d7438aa0-1e45-4112-9af4-a584825d29df" (UID: "d7438aa0-1e45-4112-9af4-a584825d29df"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.280885 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "d7438aa0-1e45-4112-9af4-a584825d29df" (UID: "d7438aa0-1e45-4112-9af4-a584825d29df"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.282168 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d7438aa0-1e45-4112-9af4-a584825d29df","Type":"ContainerDied","Data":"2492283905a5c55c900b3eba6a0122b3f3c9e6da2f0de8d0cfa8fcd7d80610e9"} Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.282233 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.282255 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2492283905a5c55c900b3eba6a0122b3f3c9e6da2f0de8d0cfa8fcd7d80610e9" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.305111 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7438aa0-1e45-4112-9af4-a584825d29df-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "d7438aa0-1e45-4112-9af4-a584825d29df" (UID: "d7438aa0-1e45-4112-9af4-a584825d29df"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.339204 4739 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.339241 4739 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d7438aa0-1e45-4112-9af4-a584825d29df-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.339261 4739 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.339275 4739 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.339288 4739 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d7438aa0-1e45-4112-9af4-a584825d29df-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.339299 4739 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d7438aa0-1e45-4112-9af4-a584825d29df-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.339309 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvzjb\" (UniqueName: \"kubernetes.io/projected/d7438aa0-1e45-4112-9af4-a584825d29df-kube-api-access-cvzjb\") on node \"crc\" DevicePath \"\"" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.360212 4739 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 01 16:51:37 crc kubenswrapper[4739]: I1201 16:51:37.441291 4739 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 01 16:51:49 crc kubenswrapper[4739]: I1201 16:51:49.927815 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 01 16:51:49 crc kubenswrapper[4739]: E1201 16:51:49.928899 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55dad234-160d-42ec-b1b1-0d21d262c1e8" containerName="extract-content" Dec 01 16:51:49 crc kubenswrapper[4739]: I1201 16:51:49.928937 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="55dad234-160d-42ec-b1b1-0d21d262c1e8" containerName="extract-content" Dec 01 16:51:49 crc kubenswrapper[4739]: E1201 16:51:49.928952 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55dad234-160d-42ec-b1b1-0d21d262c1e8" containerName="registry-server" Dec 01 16:51:49 crc kubenswrapper[4739]: I1201 16:51:49.928959 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="55dad234-160d-42ec-b1b1-0d21d262c1e8" containerName="registry-server" Dec 01 16:51:49 crc kubenswrapper[4739]: E1201 16:51:49.928976 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7438aa0-1e45-4112-9af4-a584825d29df" containerName="tempest-tests-tempest-tests-runner" Dec 01 16:51:49 crc kubenswrapper[4739]: I1201 16:51:49.928984 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7438aa0-1e45-4112-9af4-a584825d29df" containerName="tempest-tests-tempest-tests-runner" Dec 01 16:51:49 crc kubenswrapper[4739]: E1201 16:51:49.929025 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55dad234-160d-42ec-b1b1-0d21d262c1e8" containerName="extract-utilities" Dec 01 16:51:49 crc kubenswrapper[4739]: I1201 16:51:49.929034 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="55dad234-160d-42ec-b1b1-0d21d262c1e8" containerName="extract-utilities" Dec 01 16:51:49 crc kubenswrapper[4739]: I1201 16:51:49.929296 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="55dad234-160d-42ec-b1b1-0d21d262c1e8" containerName="registry-server" Dec 01 16:51:49 crc kubenswrapper[4739]: I1201 16:51:49.929318 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7438aa0-1e45-4112-9af4-a584825d29df" containerName="tempest-tests-tempest-tests-runner" Dec 01 16:51:49 crc kubenswrapper[4739]: I1201 16:51:49.930079 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 16:51:49 crc kubenswrapper[4739]: I1201 16:51:49.933534 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cpqn6" Dec 01 16:51:49 crc kubenswrapper[4739]: I1201 16:51:49.942271 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 01 16:51:49 crc kubenswrapper[4739]: I1201 16:51:49.999152 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b78ab401-a093-4c5d-aaa2-4c95534c3071\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 16:51:49 crc kubenswrapper[4739]: I1201 16:51:49.999202 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zgj9\" (UniqueName: \"kubernetes.io/projected/b78ab401-a093-4c5d-aaa2-4c95534c3071-kube-api-access-6zgj9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b78ab401-a093-4c5d-aaa2-4c95534c3071\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 16:51:50 crc kubenswrapper[4739]: I1201 16:51:50.101010 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b78ab401-a093-4c5d-aaa2-4c95534c3071\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 16:51:50 crc kubenswrapper[4739]: I1201 16:51:50.101077 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zgj9\" (UniqueName: \"kubernetes.io/projected/b78ab401-a093-4c5d-aaa2-4c95534c3071-kube-api-access-6zgj9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b78ab401-a093-4c5d-aaa2-4c95534c3071\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 16:51:50 crc kubenswrapper[4739]: I1201 16:51:50.101910 4739 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b78ab401-a093-4c5d-aaa2-4c95534c3071\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 16:51:50 crc kubenswrapper[4739]: I1201 16:51:50.124757 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zgj9\" (UniqueName: \"kubernetes.io/projected/b78ab401-a093-4c5d-aaa2-4c95534c3071-kube-api-access-6zgj9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b78ab401-a093-4c5d-aaa2-4c95534c3071\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 16:51:50 crc kubenswrapper[4739]: I1201 16:51:50.137439 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b78ab401-a093-4c5d-aaa2-4c95534c3071\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 16:51:50 crc kubenswrapper[4739]: I1201 16:51:50.253958 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 16:51:50 crc kubenswrapper[4739]: I1201 16:51:50.737876 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 01 16:51:51 crc kubenswrapper[4739]: I1201 16:51:51.414698 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b78ab401-a093-4c5d-aaa2-4c95534c3071","Type":"ContainerStarted","Data":"0698c109369a2b388b779fa37dd0d242854939f902576b409ee07d2ea4ed6b67"} Dec 01 16:51:52 crc kubenswrapper[4739]: I1201 16:51:52.424992 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b78ab401-a093-4c5d-aaa2-4c95534c3071","Type":"ContainerStarted","Data":"f4892a8f22a9533a8cee542b305519a6dfa4458a57d33b1fd9a6208c65f17663"} Dec 01 16:51:52 crc kubenswrapper[4739]: I1201 16:51:52.442015 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.641130461 podStartE2EDuration="3.441988989s" podCreationTimestamp="2025-12-01 16:51:49 +0000 UTC" firstStartedPulling="2025-12-01 16:51:50.966133341 +0000 UTC m=+4612.791879435" lastFinishedPulling="2025-12-01 16:51:51.766991869 +0000 UTC m=+4613.592737963" observedRunningTime="2025-12-01 16:51:52.436812979 +0000 UTC m=+4614.262559073" watchObservedRunningTime="2025-12-01 16:51:52.441988989 +0000 UTC m=+4614.267735103" Dec 01 16:52:33 crc kubenswrapper[4739]: I1201 16:52:33.241551 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bsh7j/must-gather-ssxt2"] Dec 01 16:52:33 crc kubenswrapper[4739]: I1201 16:52:33.243498 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/must-gather-ssxt2" Dec 01 16:52:33 crc kubenswrapper[4739]: I1201 16:52:33.245721 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bsh7j"/"openshift-service-ca.crt" Dec 01 16:52:33 crc kubenswrapper[4739]: I1201 16:52:33.248188 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-bsh7j"/"default-dockercfg-jf58g" Dec 01 16:52:33 crc kubenswrapper[4739]: I1201 16:52:33.248508 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bsh7j"/"kube-root-ca.crt" Dec 01 16:52:33 crc kubenswrapper[4739]: I1201 16:52:33.253483 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bsh7j/must-gather-ssxt2"] Dec 01 16:52:33 crc kubenswrapper[4739]: I1201 16:52:33.330391 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m49hq\" (UniqueName: \"kubernetes.io/projected/efcb363f-0055-45d3-960a-bb5eef4a7d37-kube-api-access-m49hq\") pod \"must-gather-ssxt2\" (UID: \"efcb363f-0055-45d3-960a-bb5eef4a7d37\") " pod="openshift-must-gather-bsh7j/must-gather-ssxt2" Dec 01 16:52:33 crc kubenswrapper[4739]: I1201 16:52:33.330519 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/efcb363f-0055-45d3-960a-bb5eef4a7d37-must-gather-output\") pod \"must-gather-ssxt2\" (UID: \"efcb363f-0055-45d3-960a-bb5eef4a7d37\") " pod="openshift-must-gather-bsh7j/must-gather-ssxt2" Dec 01 16:52:33 crc kubenswrapper[4739]: I1201 16:52:33.431840 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m49hq\" (UniqueName: \"kubernetes.io/projected/efcb363f-0055-45d3-960a-bb5eef4a7d37-kube-api-access-m49hq\") pod \"must-gather-ssxt2\" (UID: \"efcb363f-0055-45d3-960a-bb5eef4a7d37\") " pod="openshift-must-gather-bsh7j/must-gather-ssxt2" Dec 01 16:52:33 crc kubenswrapper[4739]: I1201 16:52:33.431915 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/efcb363f-0055-45d3-960a-bb5eef4a7d37-must-gather-output\") pod \"must-gather-ssxt2\" (UID: \"efcb363f-0055-45d3-960a-bb5eef4a7d37\") " pod="openshift-must-gather-bsh7j/must-gather-ssxt2" Dec 01 16:52:33 crc kubenswrapper[4739]: I1201 16:52:33.432524 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/efcb363f-0055-45d3-960a-bb5eef4a7d37-must-gather-output\") pod \"must-gather-ssxt2\" (UID: \"efcb363f-0055-45d3-960a-bb5eef4a7d37\") " pod="openshift-must-gather-bsh7j/must-gather-ssxt2" Dec 01 16:52:33 crc kubenswrapper[4739]: I1201 16:52:33.452304 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m49hq\" (UniqueName: \"kubernetes.io/projected/efcb363f-0055-45d3-960a-bb5eef4a7d37-kube-api-access-m49hq\") pod \"must-gather-ssxt2\" (UID: \"efcb363f-0055-45d3-960a-bb5eef4a7d37\") " pod="openshift-must-gather-bsh7j/must-gather-ssxt2" Dec 01 16:52:33 crc kubenswrapper[4739]: I1201 16:52:33.562674 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/must-gather-ssxt2" Dec 01 16:52:34 crc kubenswrapper[4739]: I1201 16:52:34.008955 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bsh7j/must-gather-ssxt2"] Dec 01 16:52:34 crc kubenswrapper[4739]: I1201 16:52:34.864780 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bsh7j/must-gather-ssxt2" event={"ID":"efcb363f-0055-45d3-960a-bb5eef4a7d37","Type":"ContainerStarted","Data":"eb552bc57a850fedaebb5b8788b905015c558d914dafdd77b4ee92428e85b9e2"} Dec 01 16:52:38 crc kubenswrapper[4739]: I1201 16:52:38.903126 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bsh7j/must-gather-ssxt2" event={"ID":"efcb363f-0055-45d3-960a-bb5eef4a7d37","Type":"ContainerStarted","Data":"92cd6803f38c3c5f14919d50392dadf5dad8102307d1caf7aff76d5976f7fa50"} Dec 01 16:52:39 crc kubenswrapper[4739]: I1201 16:52:39.913495 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bsh7j/must-gather-ssxt2" event={"ID":"efcb363f-0055-45d3-960a-bb5eef4a7d37","Type":"ContainerStarted","Data":"d8520cd9036bee74c67a88a53baf5c781fea56f100b2745a1800d9a0f2ad770b"} Dec 01 16:52:39 crc kubenswrapper[4739]: I1201 16:52:39.934278 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bsh7j/must-gather-ssxt2" podStartSLOduration=3.023017081 podStartE2EDuration="6.934252905s" podCreationTimestamp="2025-12-01 16:52:33 +0000 UTC" firstStartedPulling="2025-12-01 16:52:34.018287207 +0000 UTC m=+4655.844033321" lastFinishedPulling="2025-12-01 16:52:37.929523051 +0000 UTC m=+4659.755269145" observedRunningTime="2025-12-01 16:52:39.932119319 +0000 UTC m=+4661.757865423" watchObservedRunningTime="2025-12-01 16:52:39.934252905 +0000 UTC m=+4661.759999039" Dec 01 16:52:43 crc kubenswrapper[4739]: I1201 16:52:43.518532 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bsh7j/crc-debug-sl76c"] Dec 01 16:52:43 crc kubenswrapper[4739]: I1201 16:52:43.521459 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/crc-debug-sl76c" Dec 01 16:52:43 crc kubenswrapper[4739]: I1201 16:52:43.657524 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzckb\" (UniqueName: \"kubernetes.io/projected/1b845fb3-64e7-4372-b2ce-fdfcb874b2dd-kube-api-access-dzckb\") pod \"crc-debug-sl76c\" (UID: \"1b845fb3-64e7-4372-b2ce-fdfcb874b2dd\") " pod="openshift-must-gather-bsh7j/crc-debug-sl76c" Dec 01 16:52:43 crc kubenswrapper[4739]: I1201 16:52:43.657669 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1b845fb3-64e7-4372-b2ce-fdfcb874b2dd-host\") pod \"crc-debug-sl76c\" (UID: \"1b845fb3-64e7-4372-b2ce-fdfcb874b2dd\") " pod="openshift-must-gather-bsh7j/crc-debug-sl76c" Dec 01 16:52:43 crc kubenswrapper[4739]: I1201 16:52:43.759641 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzckb\" (UniqueName: \"kubernetes.io/projected/1b845fb3-64e7-4372-b2ce-fdfcb874b2dd-kube-api-access-dzckb\") pod \"crc-debug-sl76c\" (UID: \"1b845fb3-64e7-4372-b2ce-fdfcb874b2dd\") " pod="openshift-must-gather-bsh7j/crc-debug-sl76c" Dec 01 16:52:43 crc kubenswrapper[4739]: I1201 16:52:43.759783 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1b845fb3-64e7-4372-b2ce-fdfcb874b2dd-host\") pod \"crc-debug-sl76c\" (UID: \"1b845fb3-64e7-4372-b2ce-fdfcb874b2dd\") " pod="openshift-must-gather-bsh7j/crc-debug-sl76c" Dec 01 16:52:43 crc kubenswrapper[4739]: I1201 16:52:43.759912 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1b845fb3-64e7-4372-b2ce-fdfcb874b2dd-host\") pod \"crc-debug-sl76c\" (UID: \"1b845fb3-64e7-4372-b2ce-fdfcb874b2dd\") " pod="openshift-must-gather-bsh7j/crc-debug-sl76c" Dec 01 16:52:43 crc kubenswrapper[4739]: I1201 16:52:43.790029 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzckb\" (UniqueName: \"kubernetes.io/projected/1b845fb3-64e7-4372-b2ce-fdfcb874b2dd-kube-api-access-dzckb\") pod \"crc-debug-sl76c\" (UID: \"1b845fb3-64e7-4372-b2ce-fdfcb874b2dd\") " pod="openshift-must-gather-bsh7j/crc-debug-sl76c" Dec 01 16:52:43 crc kubenswrapper[4739]: I1201 16:52:43.841344 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/crc-debug-sl76c" Dec 01 16:52:43 crc kubenswrapper[4739]: W1201 16:52:43.881450 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b845fb3_64e7_4372_b2ce_fdfcb874b2dd.slice/crio-a4c77a74d9e3b8a93848755fd788238707b29a5aab02b856b52b23365ea6a873 WatchSource:0}: Error finding container a4c77a74d9e3b8a93848755fd788238707b29a5aab02b856b52b23365ea6a873: Status 404 returned error can't find the container with id a4c77a74d9e3b8a93848755fd788238707b29a5aab02b856b52b23365ea6a873 Dec 01 16:52:43 crc kubenswrapper[4739]: I1201 16:52:43.966921 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bsh7j/crc-debug-sl76c" event={"ID":"1b845fb3-64e7-4372-b2ce-fdfcb874b2dd","Type":"ContainerStarted","Data":"a4c77a74d9e3b8a93848755fd788238707b29a5aab02b856b52b23365ea6a873"} Dec 01 16:52:56 crc kubenswrapper[4739]: I1201 16:52:56.082359 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bsh7j/crc-debug-sl76c" event={"ID":"1b845fb3-64e7-4372-b2ce-fdfcb874b2dd","Type":"ContainerStarted","Data":"f2d0c52fc771e587715191f09ad4f8d319f245e81bb62aed0007093136bd453d"} Dec 01 16:52:56 crc kubenswrapper[4739]: I1201 16:52:56.105758 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bsh7j/crc-debug-sl76c" podStartSLOduration=1.589189042 podStartE2EDuration="13.105738696s" podCreationTimestamp="2025-12-01 16:52:43 +0000 UTC" firstStartedPulling="2025-12-01 16:52:43.883910786 +0000 UTC m=+4665.709656870" lastFinishedPulling="2025-12-01 16:52:55.40046043 +0000 UTC m=+4677.226206524" observedRunningTime="2025-12-01 16:52:56.105285992 +0000 UTC m=+4677.931032086" watchObservedRunningTime="2025-12-01 16:52:56.105738696 +0000 UTC m=+4677.931484790" Dec 01 16:53:09 crc kubenswrapper[4739]: I1201 16:53:09.622271 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:53:09 crc kubenswrapper[4739]: I1201 16:53:09.623739 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:53:32 crc kubenswrapper[4739]: I1201 16:53:32.229973 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fjss5"] Dec 01 16:53:32 crc kubenswrapper[4739]: I1201 16:53:32.232962 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:32 crc kubenswrapper[4739]: I1201 16:53:32.246035 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fjss5"] Dec 01 16:53:32 crc kubenswrapper[4739]: I1201 16:53:32.336631 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-catalog-content\") pod \"redhat-operators-fjss5\" (UID: \"a2884e85-d02c-4a19-bbe5-97f2ba4003dc\") " pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:32 crc kubenswrapper[4739]: I1201 16:53:32.336676 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-utilities\") pod \"redhat-operators-fjss5\" (UID: \"a2884e85-d02c-4a19-bbe5-97f2ba4003dc\") " pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:32 crc kubenswrapper[4739]: I1201 16:53:32.337013 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9xdz\" (UniqueName: \"kubernetes.io/projected/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-kube-api-access-k9xdz\") pod \"redhat-operators-fjss5\" (UID: \"a2884e85-d02c-4a19-bbe5-97f2ba4003dc\") " pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:32 crc kubenswrapper[4739]: I1201 16:53:32.438763 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9xdz\" (UniqueName: \"kubernetes.io/projected/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-kube-api-access-k9xdz\") pod \"redhat-operators-fjss5\" (UID: \"a2884e85-d02c-4a19-bbe5-97f2ba4003dc\") " pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:32 crc kubenswrapper[4739]: I1201 16:53:32.438851 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-catalog-content\") pod \"redhat-operators-fjss5\" (UID: \"a2884e85-d02c-4a19-bbe5-97f2ba4003dc\") " pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:32 crc kubenswrapper[4739]: I1201 16:53:32.438878 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-utilities\") pod \"redhat-operators-fjss5\" (UID: \"a2884e85-d02c-4a19-bbe5-97f2ba4003dc\") " pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:32 crc kubenswrapper[4739]: I1201 16:53:32.439360 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-utilities\") pod \"redhat-operators-fjss5\" (UID: \"a2884e85-d02c-4a19-bbe5-97f2ba4003dc\") " pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:32 crc kubenswrapper[4739]: I1201 16:53:32.439510 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-catalog-content\") pod \"redhat-operators-fjss5\" (UID: \"a2884e85-d02c-4a19-bbe5-97f2ba4003dc\") " pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:32 crc kubenswrapper[4739]: I1201 16:53:32.466880 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9xdz\" (UniqueName: \"kubernetes.io/projected/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-kube-api-access-k9xdz\") pod \"redhat-operators-fjss5\" (UID: \"a2884e85-d02c-4a19-bbe5-97f2ba4003dc\") " pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:32 crc kubenswrapper[4739]: I1201 16:53:32.589131 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:33 crc kubenswrapper[4739]: I1201 16:53:33.115941 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fjss5"] Dec 01 16:53:33 crc kubenswrapper[4739]: I1201 16:53:33.427533 4739 generic.go:334] "Generic (PLEG): container finished" podID="a2884e85-d02c-4a19-bbe5-97f2ba4003dc" containerID="0284af2b4f0e25f5953df0bf39fe040e6e0f1adf7e2736fe69fef58dff9494ab" exitCode=0 Dec 01 16:53:33 crc kubenswrapper[4739]: I1201 16:53:33.427651 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjss5" event={"ID":"a2884e85-d02c-4a19-bbe5-97f2ba4003dc","Type":"ContainerDied","Data":"0284af2b4f0e25f5953df0bf39fe040e6e0f1adf7e2736fe69fef58dff9494ab"} Dec 01 16:53:33 crc kubenswrapper[4739]: I1201 16:53:33.427850 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjss5" event={"ID":"a2884e85-d02c-4a19-bbe5-97f2ba4003dc","Type":"ContainerStarted","Data":"575a325c601ed8563a2597d16a4f8d012810ccad875aa3ec11bdad2291454c73"} Dec 01 16:53:34 crc kubenswrapper[4739]: I1201 16:53:34.438179 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjss5" event={"ID":"a2884e85-d02c-4a19-bbe5-97f2ba4003dc","Type":"ContainerStarted","Data":"c63b0b063042353e7b1fd451dc1bd446ad8ce0f06a24289fe3f48ea374bc6c2f"} Dec 01 16:53:36 crc kubenswrapper[4739]: I1201 16:53:36.467137 4739 generic.go:334] "Generic (PLEG): container finished" podID="a2884e85-d02c-4a19-bbe5-97f2ba4003dc" containerID="c63b0b063042353e7b1fd451dc1bd446ad8ce0f06a24289fe3f48ea374bc6c2f" exitCode=0 Dec 01 16:53:36 crc kubenswrapper[4739]: I1201 16:53:36.467238 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjss5" event={"ID":"a2884e85-d02c-4a19-bbe5-97f2ba4003dc","Type":"ContainerDied","Data":"c63b0b063042353e7b1fd451dc1bd446ad8ce0f06a24289fe3f48ea374bc6c2f"} Dec 01 16:53:39 crc kubenswrapper[4739]: I1201 16:53:39.498322 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjss5" event={"ID":"a2884e85-d02c-4a19-bbe5-97f2ba4003dc","Type":"ContainerStarted","Data":"5b8f85fe0e040c803af19b2050382947448481e6e317a903e17e90e77fbc6489"} Dec 01 16:53:39 crc kubenswrapper[4739]: I1201 16:53:39.527129 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fjss5" podStartSLOduration=2.453180759 podStartE2EDuration="7.527105206s" podCreationTimestamp="2025-12-01 16:53:32 +0000 UTC" firstStartedPulling="2025-12-01 16:53:33.429349143 +0000 UTC m=+4715.255095237" lastFinishedPulling="2025-12-01 16:53:38.50327359 +0000 UTC m=+4720.329019684" observedRunningTime="2025-12-01 16:53:39.517099506 +0000 UTC m=+4721.342845600" watchObservedRunningTime="2025-12-01 16:53:39.527105206 +0000 UTC m=+4721.352851300" Dec 01 16:53:39 crc kubenswrapper[4739]: I1201 16:53:39.621304 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:53:39 crc kubenswrapper[4739]: I1201 16:53:39.621353 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:53:40 crc kubenswrapper[4739]: I1201 16:53:40.510979 4739 generic.go:334] "Generic (PLEG): container finished" podID="1b845fb3-64e7-4372-b2ce-fdfcb874b2dd" containerID="f2d0c52fc771e587715191f09ad4f8d319f245e81bb62aed0007093136bd453d" exitCode=0 Dec 01 16:53:40 crc kubenswrapper[4739]: I1201 16:53:40.511174 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bsh7j/crc-debug-sl76c" event={"ID":"1b845fb3-64e7-4372-b2ce-fdfcb874b2dd","Type":"ContainerDied","Data":"f2d0c52fc771e587715191f09ad4f8d319f245e81bb62aed0007093136bd453d"} Dec 01 16:53:42 crc kubenswrapper[4739]: I1201 16:53:42.105803 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/crc-debug-sl76c" Dec 01 16:53:42 crc kubenswrapper[4739]: I1201 16:53:42.139524 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bsh7j/crc-debug-sl76c"] Dec 01 16:53:42 crc kubenswrapper[4739]: I1201 16:53:42.148904 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bsh7j/crc-debug-sl76c"] Dec 01 16:53:42 crc kubenswrapper[4739]: I1201 16:53:42.253851 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzckb\" (UniqueName: \"kubernetes.io/projected/1b845fb3-64e7-4372-b2ce-fdfcb874b2dd-kube-api-access-dzckb\") pod \"1b845fb3-64e7-4372-b2ce-fdfcb874b2dd\" (UID: \"1b845fb3-64e7-4372-b2ce-fdfcb874b2dd\") " Dec 01 16:53:42 crc kubenswrapper[4739]: I1201 16:53:42.253924 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1b845fb3-64e7-4372-b2ce-fdfcb874b2dd-host\") pod \"1b845fb3-64e7-4372-b2ce-fdfcb874b2dd\" (UID: \"1b845fb3-64e7-4372-b2ce-fdfcb874b2dd\") " Dec 01 16:53:42 crc kubenswrapper[4739]: I1201 16:53:42.254074 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1b845fb3-64e7-4372-b2ce-fdfcb874b2dd-host" (OuterVolumeSpecName: "host") pod "1b845fb3-64e7-4372-b2ce-fdfcb874b2dd" (UID: "1b845fb3-64e7-4372-b2ce-fdfcb874b2dd"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:53:42 crc kubenswrapper[4739]: I1201 16:53:42.254443 4739 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1b845fb3-64e7-4372-b2ce-fdfcb874b2dd-host\") on node \"crc\" DevicePath \"\"" Dec 01 16:53:42 crc kubenswrapper[4739]: I1201 16:53:42.260982 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b845fb3-64e7-4372-b2ce-fdfcb874b2dd-kube-api-access-dzckb" (OuterVolumeSpecName: "kube-api-access-dzckb") pod "1b845fb3-64e7-4372-b2ce-fdfcb874b2dd" (UID: "1b845fb3-64e7-4372-b2ce-fdfcb874b2dd"). InnerVolumeSpecName "kube-api-access-dzckb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:53:42 crc kubenswrapper[4739]: I1201 16:53:42.356719 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzckb\" (UniqueName: \"kubernetes.io/projected/1b845fb3-64e7-4372-b2ce-fdfcb874b2dd-kube-api-access-dzckb\") on node \"crc\" DevicePath \"\"" Dec 01 16:53:42 crc kubenswrapper[4739]: I1201 16:53:42.488299 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b845fb3-64e7-4372-b2ce-fdfcb874b2dd" path="/var/lib/kubelet/pods/1b845fb3-64e7-4372-b2ce-fdfcb874b2dd/volumes" Dec 01 16:53:42 crc kubenswrapper[4739]: I1201 16:53:42.527631 4739 scope.go:117] "RemoveContainer" containerID="f2d0c52fc771e587715191f09ad4f8d319f245e81bb62aed0007093136bd453d" Dec 01 16:53:42 crc kubenswrapper[4739]: I1201 16:53:42.527702 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/crc-debug-sl76c" Dec 01 16:53:42 crc kubenswrapper[4739]: I1201 16:53:42.589660 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:42 crc kubenswrapper[4739]: I1201 16:53:42.590051 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:43 crc kubenswrapper[4739]: I1201 16:53:43.333397 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bsh7j/crc-debug-mw59f"] Dec 01 16:53:43 crc kubenswrapper[4739]: E1201 16:53:43.334266 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b845fb3-64e7-4372-b2ce-fdfcb874b2dd" containerName="container-00" Dec 01 16:53:43 crc kubenswrapper[4739]: I1201 16:53:43.334285 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b845fb3-64e7-4372-b2ce-fdfcb874b2dd" containerName="container-00" Dec 01 16:53:43 crc kubenswrapper[4739]: I1201 16:53:43.334529 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b845fb3-64e7-4372-b2ce-fdfcb874b2dd" containerName="container-00" Dec 01 16:53:43 crc kubenswrapper[4739]: I1201 16:53:43.335296 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/crc-debug-mw59f" Dec 01 16:53:43 crc kubenswrapper[4739]: I1201 16:53:43.481547 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmqhd\" (UniqueName: \"kubernetes.io/projected/78df3f14-d7a1-4311-8158-cd1d88710ba2-kube-api-access-dmqhd\") pod \"crc-debug-mw59f\" (UID: \"78df3f14-d7a1-4311-8158-cd1d88710ba2\") " pod="openshift-must-gather-bsh7j/crc-debug-mw59f" Dec 01 16:53:43 crc kubenswrapper[4739]: I1201 16:53:43.481826 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/78df3f14-d7a1-4311-8158-cd1d88710ba2-host\") pod \"crc-debug-mw59f\" (UID: \"78df3f14-d7a1-4311-8158-cd1d88710ba2\") " pod="openshift-must-gather-bsh7j/crc-debug-mw59f" Dec 01 16:53:43 crc kubenswrapper[4739]: I1201 16:53:43.586610 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/78df3f14-d7a1-4311-8158-cd1d88710ba2-host\") pod \"crc-debug-mw59f\" (UID: \"78df3f14-d7a1-4311-8158-cd1d88710ba2\") " pod="openshift-must-gather-bsh7j/crc-debug-mw59f" Dec 01 16:53:43 crc kubenswrapper[4739]: I1201 16:53:43.586981 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/78df3f14-d7a1-4311-8158-cd1d88710ba2-host\") pod \"crc-debug-mw59f\" (UID: \"78df3f14-d7a1-4311-8158-cd1d88710ba2\") " pod="openshift-must-gather-bsh7j/crc-debug-mw59f" Dec 01 16:53:43 crc kubenswrapper[4739]: I1201 16:53:43.590506 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmqhd\" (UniqueName: \"kubernetes.io/projected/78df3f14-d7a1-4311-8158-cd1d88710ba2-kube-api-access-dmqhd\") pod \"crc-debug-mw59f\" (UID: \"78df3f14-d7a1-4311-8158-cd1d88710ba2\") " pod="openshift-must-gather-bsh7j/crc-debug-mw59f" Dec 01 16:53:43 crc kubenswrapper[4739]: I1201 16:53:43.613072 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmqhd\" (UniqueName: \"kubernetes.io/projected/78df3f14-d7a1-4311-8158-cd1d88710ba2-kube-api-access-dmqhd\") pod \"crc-debug-mw59f\" (UID: \"78df3f14-d7a1-4311-8158-cd1d88710ba2\") " pod="openshift-must-gather-bsh7j/crc-debug-mw59f" Dec 01 16:53:43 crc kubenswrapper[4739]: I1201 16:53:43.644521 4739 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fjss5" podUID="a2884e85-d02c-4a19-bbe5-97f2ba4003dc" containerName="registry-server" probeResult="failure" output=< Dec 01 16:53:43 crc kubenswrapper[4739]: timeout: failed to connect service ":50051" within 1s Dec 01 16:53:43 crc kubenswrapper[4739]: > Dec 01 16:53:43 crc kubenswrapper[4739]: I1201 16:53:43.655674 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/crc-debug-mw59f" Dec 01 16:53:44 crc kubenswrapper[4739]: I1201 16:53:44.549911 4739 generic.go:334] "Generic (PLEG): container finished" podID="78df3f14-d7a1-4311-8158-cd1d88710ba2" containerID="0ba245c11fbd431a71abab2e4668376b895d1573285b82b9f5dafb2b2e88aef7" exitCode=0 Dec 01 16:53:44 crc kubenswrapper[4739]: I1201 16:53:44.550004 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bsh7j/crc-debug-mw59f" event={"ID":"78df3f14-d7a1-4311-8158-cd1d88710ba2","Type":"ContainerDied","Data":"0ba245c11fbd431a71abab2e4668376b895d1573285b82b9f5dafb2b2e88aef7"} Dec 01 16:53:44 crc kubenswrapper[4739]: I1201 16:53:44.550297 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bsh7j/crc-debug-mw59f" event={"ID":"78df3f14-d7a1-4311-8158-cd1d88710ba2","Type":"ContainerStarted","Data":"25d3575f6b6ec8ec707369518a3866a396db042f55124298fcd39cad3b22f824"} Dec 01 16:53:45 crc kubenswrapper[4739]: I1201 16:53:45.658998 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/crc-debug-mw59f" Dec 01 16:53:45 crc kubenswrapper[4739]: I1201 16:53:45.729848 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmqhd\" (UniqueName: \"kubernetes.io/projected/78df3f14-d7a1-4311-8158-cd1d88710ba2-kube-api-access-dmqhd\") pod \"78df3f14-d7a1-4311-8158-cd1d88710ba2\" (UID: \"78df3f14-d7a1-4311-8158-cd1d88710ba2\") " Dec 01 16:53:45 crc kubenswrapper[4739]: I1201 16:53:45.729982 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/78df3f14-d7a1-4311-8158-cd1d88710ba2-host\") pod \"78df3f14-d7a1-4311-8158-cd1d88710ba2\" (UID: \"78df3f14-d7a1-4311-8158-cd1d88710ba2\") " Dec 01 16:53:45 crc kubenswrapper[4739]: I1201 16:53:45.730111 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/78df3f14-d7a1-4311-8158-cd1d88710ba2-host" (OuterVolumeSpecName: "host") pod "78df3f14-d7a1-4311-8158-cd1d88710ba2" (UID: "78df3f14-d7a1-4311-8158-cd1d88710ba2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:53:45 crc kubenswrapper[4739]: I1201 16:53:45.730505 4739 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/78df3f14-d7a1-4311-8158-cd1d88710ba2-host\") on node \"crc\" DevicePath \"\"" Dec 01 16:53:45 crc kubenswrapper[4739]: I1201 16:53:45.735770 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78df3f14-d7a1-4311-8158-cd1d88710ba2-kube-api-access-dmqhd" (OuterVolumeSpecName: "kube-api-access-dmqhd") pod "78df3f14-d7a1-4311-8158-cd1d88710ba2" (UID: "78df3f14-d7a1-4311-8158-cd1d88710ba2"). InnerVolumeSpecName "kube-api-access-dmqhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:53:45 crc kubenswrapper[4739]: I1201 16:53:45.832867 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmqhd\" (UniqueName: \"kubernetes.io/projected/78df3f14-d7a1-4311-8158-cd1d88710ba2-kube-api-access-dmqhd\") on node \"crc\" DevicePath \"\"" Dec 01 16:53:46 crc kubenswrapper[4739]: I1201 16:53:46.570985 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bsh7j/crc-debug-mw59f" event={"ID":"78df3f14-d7a1-4311-8158-cd1d88710ba2","Type":"ContainerDied","Data":"25d3575f6b6ec8ec707369518a3866a396db042f55124298fcd39cad3b22f824"} Dec 01 16:53:46 crc kubenswrapper[4739]: I1201 16:53:46.571368 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25d3575f6b6ec8ec707369518a3866a396db042f55124298fcd39cad3b22f824" Dec 01 16:53:46 crc kubenswrapper[4739]: I1201 16:53:46.571456 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/crc-debug-mw59f" Dec 01 16:53:47 crc kubenswrapper[4739]: I1201 16:53:47.199364 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bsh7j/crc-debug-mw59f"] Dec 01 16:53:47 crc kubenswrapper[4739]: I1201 16:53:47.208924 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bsh7j/crc-debug-mw59f"] Dec 01 16:53:48 crc kubenswrapper[4739]: I1201 16:53:48.490582 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78df3f14-d7a1-4311-8158-cd1d88710ba2" path="/var/lib/kubelet/pods/78df3f14-d7a1-4311-8158-cd1d88710ba2/volumes" Dec 01 16:53:48 crc kubenswrapper[4739]: I1201 16:53:48.492640 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bsh7j/crc-debug-pds5w"] Dec 01 16:53:48 crc kubenswrapper[4739]: E1201 16:53:48.493355 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78df3f14-d7a1-4311-8158-cd1d88710ba2" containerName="container-00" Dec 01 16:53:48 crc kubenswrapper[4739]: I1201 16:53:48.493383 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="78df3f14-d7a1-4311-8158-cd1d88710ba2" containerName="container-00" Dec 01 16:53:48 crc kubenswrapper[4739]: I1201 16:53:48.493642 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="78df3f14-d7a1-4311-8158-cd1d88710ba2" containerName="container-00" Dec 01 16:53:48 crc kubenswrapper[4739]: I1201 16:53:48.494469 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/crc-debug-pds5w" Dec 01 16:53:48 crc kubenswrapper[4739]: I1201 16:53:48.587341 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69kl2\" (UniqueName: \"kubernetes.io/projected/94fa0fcc-4611-420a-8f26-e363e22384c5-kube-api-access-69kl2\") pod \"crc-debug-pds5w\" (UID: \"94fa0fcc-4611-420a-8f26-e363e22384c5\") " pod="openshift-must-gather-bsh7j/crc-debug-pds5w" Dec 01 16:53:48 crc kubenswrapper[4739]: I1201 16:53:48.587808 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94fa0fcc-4611-420a-8f26-e363e22384c5-host\") pod \"crc-debug-pds5w\" (UID: \"94fa0fcc-4611-420a-8f26-e363e22384c5\") " pod="openshift-must-gather-bsh7j/crc-debug-pds5w" Dec 01 16:53:48 crc kubenswrapper[4739]: I1201 16:53:48.690532 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69kl2\" (UniqueName: \"kubernetes.io/projected/94fa0fcc-4611-420a-8f26-e363e22384c5-kube-api-access-69kl2\") pod \"crc-debug-pds5w\" (UID: \"94fa0fcc-4611-420a-8f26-e363e22384c5\") " pod="openshift-must-gather-bsh7j/crc-debug-pds5w" Dec 01 16:53:48 crc kubenswrapper[4739]: I1201 16:53:48.690744 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94fa0fcc-4611-420a-8f26-e363e22384c5-host\") pod \"crc-debug-pds5w\" (UID: \"94fa0fcc-4611-420a-8f26-e363e22384c5\") " pod="openshift-must-gather-bsh7j/crc-debug-pds5w" Dec 01 16:53:48 crc kubenswrapper[4739]: I1201 16:53:48.690942 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94fa0fcc-4611-420a-8f26-e363e22384c5-host\") pod \"crc-debug-pds5w\" (UID: \"94fa0fcc-4611-420a-8f26-e363e22384c5\") " pod="openshift-must-gather-bsh7j/crc-debug-pds5w" Dec 01 16:53:48 crc kubenswrapper[4739]: I1201 16:53:48.708483 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69kl2\" (UniqueName: \"kubernetes.io/projected/94fa0fcc-4611-420a-8f26-e363e22384c5-kube-api-access-69kl2\") pod \"crc-debug-pds5w\" (UID: \"94fa0fcc-4611-420a-8f26-e363e22384c5\") " pod="openshift-must-gather-bsh7j/crc-debug-pds5w" Dec 01 16:53:48 crc kubenswrapper[4739]: I1201 16:53:48.812377 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/crc-debug-pds5w" Dec 01 16:53:49 crc kubenswrapper[4739]: I1201 16:53:49.600802 4739 generic.go:334] "Generic (PLEG): container finished" podID="94fa0fcc-4611-420a-8f26-e363e22384c5" containerID="d7fd9fbf1ec95d8f9bcf8ddce4e286af1d764a44ffa278496b77aba939a815b6" exitCode=0 Dec 01 16:53:49 crc kubenswrapper[4739]: I1201 16:53:49.600900 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bsh7j/crc-debug-pds5w" event={"ID":"94fa0fcc-4611-420a-8f26-e363e22384c5","Type":"ContainerDied","Data":"d7fd9fbf1ec95d8f9bcf8ddce4e286af1d764a44ffa278496b77aba939a815b6"} Dec 01 16:53:49 crc kubenswrapper[4739]: I1201 16:53:49.601156 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bsh7j/crc-debug-pds5w" event={"ID":"94fa0fcc-4611-420a-8f26-e363e22384c5","Type":"ContainerStarted","Data":"e15f655b7dfd7e43389e65bfd48bc3dfa2e04e76ac467c6521a86ee3abce4d4c"} Dec 01 16:53:49 crc kubenswrapper[4739]: I1201 16:53:49.644449 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bsh7j/crc-debug-pds5w"] Dec 01 16:53:49 crc kubenswrapper[4739]: I1201 16:53:49.652949 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bsh7j/crc-debug-pds5w"] Dec 01 16:53:50 crc kubenswrapper[4739]: I1201 16:53:50.740804 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/crc-debug-pds5w" Dec 01 16:53:50 crc kubenswrapper[4739]: I1201 16:53:50.832944 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94fa0fcc-4611-420a-8f26-e363e22384c5-host\") pod \"94fa0fcc-4611-420a-8f26-e363e22384c5\" (UID: \"94fa0fcc-4611-420a-8f26-e363e22384c5\") " Dec 01 16:53:50 crc kubenswrapper[4739]: I1201 16:53:50.833178 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69kl2\" (UniqueName: \"kubernetes.io/projected/94fa0fcc-4611-420a-8f26-e363e22384c5-kube-api-access-69kl2\") pod \"94fa0fcc-4611-420a-8f26-e363e22384c5\" (UID: \"94fa0fcc-4611-420a-8f26-e363e22384c5\") " Dec 01 16:53:50 crc kubenswrapper[4739]: I1201 16:53:50.834190 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/94fa0fcc-4611-420a-8f26-e363e22384c5-host" (OuterVolumeSpecName: "host") pod "94fa0fcc-4611-420a-8f26-e363e22384c5" (UID: "94fa0fcc-4611-420a-8f26-e363e22384c5"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:53:50 crc kubenswrapper[4739]: I1201 16:53:50.840556 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94fa0fcc-4611-420a-8f26-e363e22384c5-kube-api-access-69kl2" (OuterVolumeSpecName: "kube-api-access-69kl2") pod "94fa0fcc-4611-420a-8f26-e363e22384c5" (UID: "94fa0fcc-4611-420a-8f26-e363e22384c5"). InnerVolumeSpecName "kube-api-access-69kl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:53:50 crc kubenswrapper[4739]: I1201 16:53:50.935716 4739 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94fa0fcc-4611-420a-8f26-e363e22384c5-host\") on node \"crc\" DevicePath \"\"" Dec 01 16:53:50 crc kubenswrapper[4739]: I1201 16:53:50.935754 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69kl2\" (UniqueName: \"kubernetes.io/projected/94fa0fcc-4611-420a-8f26-e363e22384c5-kube-api-access-69kl2\") on node \"crc\" DevicePath \"\"" Dec 01 16:53:51 crc kubenswrapper[4739]: I1201 16:53:51.621581 4739 scope.go:117] "RemoveContainer" containerID="d7fd9fbf1ec95d8f9bcf8ddce4e286af1d764a44ffa278496b77aba939a815b6" Dec 01 16:53:51 crc kubenswrapper[4739]: I1201 16:53:51.621651 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/crc-debug-pds5w" Dec 01 16:53:52 crc kubenswrapper[4739]: I1201 16:53:52.489402 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94fa0fcc-4611-420a-8f26-e363e22384c5" path="/var/lib/kubelet/pods/94fa0fcc-4611-420a-8f26-e363e22384c5/volumes" Dec 01 16:53:52 crc kubenswrapper[4739]: I1201 16:53:52.637972 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:52 crc kubenswrapper[4739]: I1201 16:53:52.684653 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:52 crc kubenswrapper[4739]: I1201 16:53:52.878296 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fjss5"] Dec 01 16:53:54 crc kubenswrapper[4739]: I1201 16:53:54.663738 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fjss5" podUID="a2884e85-d02c-4a19-bbe5-97f2ba4003dc" containerName="registry-server" containerID="cri-o://5b8f85fe0e040c803af19b2050382947448481e6e317a903e17e90e77fbc6489" gracePeriod=2 Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.094479 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.217716 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-utilities\") pod \"a2884e85-d02c-4a19-bbe5-97f2ba4003dc\" (UID: \"a2884e85-d02c-4a19-bbe5-97f2ba4003dc\") " Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.217861 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-catalog-content\") pod \"a2884e85-d02c-4a19-bbe5-97f2ba4003dc\" (UID: \"a2884e85-d02c-4a19-bbe5-97f2ba4003dc\") " Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.217931 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9xdz\" (UniqueName: \"kubernetes.io/projected/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-kube-api-access-k9xdz\") pod \"a2884e85-d02c-4a19-bbe5-97f2ba4003dc\" (UID: \"a2884e85-d02c-4a19-bbe5-97f2ba4003dc\") " Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.218905 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-utilities" (OuterVolumeSpecName: "utilities") pod "a2884e85-d02c-4a19-bbe5-97f2ba4003dc" (UID: "a2884e85-d02c-4a19-bbe5-97f2ba4003dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.223550 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-kube-api-access-k9xdz" (OuterVolumeSpecName: "kube-api-access-k9xdz") pod "a2884e85-d02c-4a19-bbe5-97f2ba4003dc" (UID: "a2884e85-d02c-4a19-bbe5-97f2ba4003dc"). InnerVolumeSpecName "kube-api-access-k9xdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.320024 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9xdz\" (UniqueName: \"kubernetes.io/projected/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-kube-api-access-k9xdz\") on node \"crc\" DevicePath \"\"" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.320054 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.332561 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2884e85-d02c-4a19-bbe5-97f2ba4003dc" (UID: "a2884e85-d02c-4a19-bbe5-97f2ba4003dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.421996 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2884e85-d02c-4a19-bbe5-97f2ba4003dc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.675700 4739 generic.go:334] "Generic (PLEG): container finished" podID="a2884e85-d02c-4a19-bbe5-97f2ba4003dc" containerID="5b8f85fe0e040c803af19b2050382947448481e6e317a903e17e90e77fbc6489" exitCode=0 Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.675755 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjss5" event={"ID":"a2884e85-d02c-4a19-bbe5-97f2ba4003dc","Type":"ContainerDied","Data":"5b8f85fe0e040c803af19b2050382947448481e6e317a903e17e90e77fbc6489"} Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.675804 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjss5" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.675826 4739 scope.go:117] "RemoveContainer" containerID="5b8f85fe0e040c803af19b2050382947448481e6e317a903e17e90e77fbc6489" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.675812 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjss5" event={"ID":"a2884e85-d02c-4a19-bbe5-97f2ba4003dc","Type":"ContainerDied","Data":"575a325c601ed8563a2597d16a4f8d012810ccad875aa3ec11bdad2291454c73"} Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.699005 4739 scope.go:117] "RemoveContainer" containerID="c63b0b063042353e7b1fd451dc1bd446ad8ce0f06a24289fe3f48ea374bc6c2f" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.729372 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fjss5"] Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.737011 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fjss5"] Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.748336 4739 scope.go:117] "RemoveContainer" containerID="0284af2b4f0e25f5953df0bf39fe040e6e0f1adf7e2736fe69fef58dff9494ab" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.787962 4739 scope.go:117] "RemoveContainer" containerID="5b8f85fe0e040c803af19b2050382947448481e6e317a903e17e90e77fbc6489" Dec 01 16:53:55 crc kubenswrapper[4739]: E1201 16:53:55.788484 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b8f85fe0e040c803af19b2050382947448481e6e317a903e17e90e77fbc6489\": container with ID starting with 5b8f85fe0e040c803af19b2050382947448481e6e317a903e17e90e77fbc6489 not found: ID does not exist" containerID="5b8f85fe0e040c803af19b2050382947448481e6e317a903e17e90e77fbc6489" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.788739 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b8f85fe0e040c803af19b2050382947448481e6e317a903e17e90e77fbc6489"} err="failed to get container status \"5b8f85fe0e040c803af19b2050382947448481e6e317a903e17e90e77fbc6489\": rpc error: code = NotFound desc = could not find container \"5b8f85fe0e040c803af19b2050382947448481e6e317a903e17e90e77fbc6489\": container with ID starting with 5b8f85fe0e040c803af19b2050382947448481e6e317a903e17e90e77fbc6489 not found: ID does not exist" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.788931 4739 scope.go:117] "RemoveContainer" containerID="c63b0b063042353e7b1fd451dc1bd446ad8ce0f06a24289fe3f48ea374bc6c2f" Dec 01 16:53:55 crc kubenswrapper[4739]: E1201 16:53:55.789411 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c63b0b063042353e7b1fd451dc1bd446ad8ce0f06a24289fe3f48ea374bc6c2f\": container with ID starting with c63b0b063042353e7b1fd451dc1bd446ad8ce0f06a24289fe3f48ea374bc6c2f not found: ID does not exist" containerID="c63b0b063042353e7b1fd451dc1bd446ad8ce0f06a24289fe3f48ea374bc6c2f" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.789470 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c63b0b063042353e7b1fd451dc1bd446ad8ce0f06a24289fe3f48ea374bc6c2f"} err="failed to get container status \"c63b0b063042353e7b1fd451dc1bd446ad8ce0f06a24289fe3f48ea374bc6c2f\": rpc error: code = NotFound desc = could not find container \"c63b0b063042353e7b1fd451dc1bd446ad8ce0f06a24289fe3f48ea374bc6c2f\": container with ID starting with c63b0b063042353e7b1fd451dc1bd446ad8ce0f06a24289fe3f48ea374bc6c2f not found: ID does not exist" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.789492 4739 scope.go:117] "RemoveContainer" containerID="0284af2b4f0e25f5953df0bf39fe040e6e0f1adf7e2736fe69fef58dff9494ab" Dec 01 16:53:55 crc kubenswrapper[4739]: E1201 16:53:55.790443 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0284af2b4f0e25f5953df0bf39fe040e6e0f1adf7e2736fe69fef58dff9494ab\": container with ID starting with 0284af2b4f0e25f5953df0bf39fe040e6e0f1adf7e2736fe69fef58dff9494ab not found: ID does not exist" containerID="0284af2b4f0e25f5953df0bf39fe040e6e0f1adf7e2736fe69fef58dff9494ab" Dec 01 16:53:55 crc kubenswrapper[4739]: I1201 16:53:55.790490 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0284af2b4f0e25f5953df0bf39fe040e6e0f1adf7e2736fe69fef58dff9494ab"} err="failed to get container status \"0284af2b4f0e25f5953df0bf39fe040e6e0f1adf7e2736fe69fef58dff9494ab\": rpc error: code = NotFound desc = could not find container \"0284af2b4f0e25f5953df0bf39fe040e6e0f1adf7e2736fe69fef58dff9494ab\": container with ID starting with 0284af2b4f0e25f5953df0bf39fe040e6e0f1adf7e2736fe69fef58dff9494ab not found: ID does not exist" Dec 01 16:53:56 crc kubenswrapper[4739]: I1201 16:53:56.509044 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2884e85-d02c-4a19-bbe5-97f2ba4003dc" path="/var/lib/kubelet/pods/a2884e85-d02c-4a19-bbe5-97f2ba4003dc/volumes" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.512839 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k8hjh"] Dec 01 16:54:07 crc kubenswrapper[4739]: E1201 16:54:07.513887 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2884e85-d02c-4a19-bbe5-97f2ba4003dc" containerName="extract-content" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.513904 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2884e85-d02c-4a19-bbe5-97f2ba4003dc" containerName="extract-content" Dec 01 16:54:07 crc kubenswrapper[4739]: E1201 16:54:07.513920 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94fa0fcc-4611-420a-8f26-e363e22384c5" containerName="container-00" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.513927 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="94fa0fcc-4611-420a-8f26-e363e22384c5" containerName="container-00" Dec 01 16:54:07 crc kubenswrapper[4739]: E1201 16:54:07.513953 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2884e85-d02c-4a19-bbe5-97f2ba4003dc" containerName="extract-utilities" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.513961 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2884e85-d02c-4a19-bbe5-97f2ba4003dc" containerName="extract-utilities" Dec 01 16:54:07 crc kubenswrapper[4739]: E1201 16:54:07.513984 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2884e85-d02c-4a19-bbe5-97f2ba4003dc" containerName="registry-server" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.513990 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2884e85-d02c-4a19-bbe5-97f2ba4003dc" containerName="registry-server" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.514185 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2884e85-d02c-4a19-bbe5-97f2ba4003dc" containerName="registry-server" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.514215 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="94fa0fcc-4611-420a-8f26-e363e22384c5" containerName="container-00" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.515778 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.524354 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k8hjh"] Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.569655 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19488820-f73d-4975-9444-e6efe1c5dba8-catalog-content\") pod \"certified-operators-k8hjh\" (UID: \"19488820-f73d-4975-9444-e6efe1c5dba8\") " pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.569761 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19488820-f73d-4975-9444-e6efe1c5dba8-utilities\") pod \"certified-operators-k8hjh\" (UID: \"19488820-f73d-4975-9444-e6efe1c5dba8\") " pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.569795 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99nxs\" (UniqueName: \"kubernetes.io/projected/19488820-f73d-4975-9444-e6efe1c5dba8-kube-api-access-99nxs\") pod \"certified-operators-k8hjh\" (UID: \"19488820-f73d-4975-9444-e6efe1c5dba8\") " pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.671465 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19488820-f73d-4975-9444-e6efe1c5dba8-catalog-content\") pod \"certified-operators-k8hjh\" (UID: \"19488820-f73d-4975-9444-e6efe1c5dba8\") " pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.671564 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19488820-f73d-4975-9444-e6efe1c5dba8-utilities\") pod \"certified-operators-k8hjh\" (UID: \"19488820-f73d-4975-9444-e6efe1c5dba8\") " pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.671589 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99nxs\" (UniqueName: \"kubernetes.io/projected/19488820-f73d-4975-9444-e6efe1c5dba8-kube-api-access-99nxs\") pod \"certified-operators-k8hjh\" (UID: \"19488820-f73d-4975-9444-e6efe1c5dba8\") " pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.672087 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19488820-f73d-4975-9444-e6efe1c5dba8-catalog-content\") pod \"certified-operators-k8hjh\" (UID: \"19488820-f73d-4975-9444-e6efe1c5dba8\") " pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.672111 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19488820-f73d-4975-9444-e6efe1c5dba8-utilities\") pod \"certified-operators-k8hjh\" (UID: \"19488820-f73d-4975-9444-e6efe1c5dba8\") " pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.710678 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99nxs\" (UniqueName: \"kubernetes.io/projected/19488820-f73d-4975-9444-e6efe1c5dba8-kube-api-access-99nxs\") pod \"certified-operators-k8hjh\" (UID: \"19488820-f73d-4975-9444-e6efe1c5dba8\") " pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:07 crc kubenswrapper[4739]: I1201 16:54:07.841510 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:08 crc kubenswrapper[4739]: I1201 16:54:08.340915 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k8hjh"] Dec 01 16:54:08 crc kubenswrapper[4739]: I1201 16:54:08.799874 4739 generic.go:334] "Generic (PLEG): container finished" podID="19488820-f73d-4975-9444-e6efe1c5dba8" containerID="bbf2d1483f512131b96ad6bb670f2cda85aed21ca6b3c8d8b818536f5c2096ea" exitCode=0 Dec 01 16:54:08 crc kubenswrapper[4739]: I1201 16:54:08.799939 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k8hjh" event={"ID":"19488820-f73d-4975-9444-e6efe1c5dba8","Type":"ContainerDied","Data":"bbf2d1483f512131b96ad6bb670f2cda85aed21ca6b3c8d8b818536f5c2096ea"} Dec 01 16:54:08 crc kubenswrapper[4739]: I1201 16:54:08.800259 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k8hjh" event={"ID":"19488820-f73d-4975-9444-e6efe1c5dba8","Type":"ContainerStarted","Data":"e29a558cc02d1f830786e0b79ee76606023e401350bf4d9008813a7a8140ec4a"} Dec 01 16:54:09 crc kubenswrapper[4739]: I1201 16:54:09.621896 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:54:09 crc kubenswrapper[4739]: I1201 16:54:09.624487 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:54:09 crc kubenswrapper[4739]: I1201 16:54:09.624555 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 16:54:09 crc kubenswrapper[4739]: I1201 16:54:09.625606 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af6de8887e1f92dd65adf332c6767d5ff238c9babbbeb1db75ee37509b77d751"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:54:09 crc kubenswrapper[4739]: I1201 16:54:09.625677 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://af6de8887e1f92dd65adf332c6767d5ff238c9babbbeb1db75ee37509b77d751" gracePeriod=600 Dec 01 16:54:09 crc kubenswrapper[4739]: I1201 16:54:09.816601 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="af6de8887e1f92dd65adf332c6767d5ff238c9babbbeb1db75ee37509b77d751" exitCode=0 Dec 01 16:54:09 crc kubenswrapper[4739]: I1201 16:54:09.816670 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"af6de8887e1f92dd65adf332c6767d5ff238c9babbbeb1db75ee37509b77d751"} Dec 01 16:54:09 crc kubenswrapper[4739]: I1201 16:54:09.816706 4739 scope.go:117] "RemoveContainer" containerID="2de8ed42fbbf4df1b368d070e2cd78636e562de5ea379b82c92991ae04968a59" Dec 01 16:54:10 crc kubenswrapper[4739]: I1201 16:54:10.826762 4739 generic.go:334] "Generic (PLEG): container finished" podID="19488820-f73d-4975-9444-e6efe1c5dba8" containerID="09956881a68a2123539f0fcd20c25975c5074dfd44f181b1c88346ce90696b49" exitCode=0 Dec 01 16:54:10 crc kubenswrapper[4739]: I1201 16:54:10.826886 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k8hjh" event={"ID":"19488820-f73d-4975-9444-e6efe1c5dba8","Type":"ContainerDied","Data":"09956881a68a2123539f0fcd20c25975c5074dfd44f181b1c88346ce90696b49"} Dec 01 16:54:10 crc kubenswrapper[4739]: I1201 16:54:10.835628 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6"} Dec 01 16:54:11 crc kubenswrapper[4739]: I1201 16:54:11.851565 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k8hjh" event={"ID":"19488820-f73d-4975-9444-e6efe1c5dba8","Type":"ContainerStarted","Data":"ff887759f80f5295b52f810d14bd483d3d05170a2a66591a4cc3d8544afd5eb8"} Dec 01 16:54:11 crc kubenswrapper[4739]: I1201 16:54:11.874138 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k8hjh" podStartSLOduration=2.363968165 podStartE2EDuration="4.87411306s" podCreationTimestamp="2025-12-01 16:54:07 +0000 UTC" firstStartedPulling="2025-12-01 16:54:08.801919324 +0000 UTC m=+4750.627665418" lastFinishedPulling="2025-12-01 16:54:11.312064219 +0000 UTC m=+4753.137810313" observedRunningTime="2025-12-01 16:54:11.872885433 +0000 UTC m=+4753.698631547" watchObservedRunningTime="2025-12-01 16:54:11.87411306 +0000 UTC m=+4753.699859164" Dec 01 16:54:17 crc kubenswrapper[4739]: I1201 16:54:17.842985 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:17 crc kubenswrapper[4739]: I1201 16:54:17.843706 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:17 crc kubenswrapper[4739]: I1201 16:54:17.895460 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:17 crc kubenswrapper[4739]: I1201 16:54:17.976461 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:18 crc kubenswrapper[4739]: I1201 16:54:18.137542 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k8hjh"] Dec 01 16:54:19 crc kubenswrapper[4739]: I1201 16:54:19.930126 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k8hjh" podUID="19488820-f73d-4975-9444-e6efe1c5dba8" containerName="registry-server" containerID="cri-o://ff887759f80f5295b52f810d14bd483d3d05170a2a66591a4cc3d8544afd5eb8" gracePeriod=2 Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.528311 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.643324 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99nxs\" (UniqueName: \"kubernetes.io/projected/19488820-f73d-4975-9444-e6efe1c5dba8-kube-api-access-99nxs\") pod \"19488820-f73d-4975-9444-e6efe1c5dba8\" (UID: \"19488820-f73d-4975-9444-e6efe1c5dba8\") " Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.643619 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19488820-f73d-4975-9444-e6efe1c5dba8-catalog-content\") pod \"19488820-f73d-4975-9444-e6efe1c5dba8\" (UID: \"19488820-f73d-4975-9444-e6efe1c5dba8\") " Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.643686 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19488820-f73d-4975-9444-e6efe1c5dba8-utilities\") pod \"19488820-f73d-4975-9444-e6efe1c5dba8\" (UID: \"19488820-f73d-4975-9444-e6efe1c5dba8\") " Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.644505 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19488820-f73d-4975-9444-e6efe1c5dba8-utilities" (OuterVolumeSpecName: "utilities") pod "19488820-f73d-4975-9444-e6efe1c5dba8" (UID: "19488820-f73d-4975-9444-e6efe1c5dba8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.651137 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19488820-f73d-4975-9444-e6efe1c5dba8-kube-api-access-99nxs" (OuterVolumeSpecName: "kube-api-access-99nxs") pod "19488820-f73d-4975-9444-e6efe1c5dba8" (UID: "19488820-f73d-4975-9444-e6efe1c5dba8"). InnerVolumeSpecName "kube-api-access-99nxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.708364 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19488820-f73d-4975-9444-e6efe1c5dba8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19488820-f73d-4975-9444-e6efe1c5dba8" (UID: "19488820-f73d-4975-9444-e6efe1c5dba8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.746194 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19488820-f73d-4975-9444-e6efe1c5dba8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.746238 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19488820-f73d-4975-9444-e6efe1c5dba8-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.746252 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99nxs\" (UniqueName: \"kubernetes.io/projected/19488820-f73d-4975-9444-e6efe1c5dba8-kube-api-access-99nxs\") on node \"crc\" DevicePath \"\"" Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.947991 4739 generic.go:334] "Generic (PLEG): container finished" podID="19488820-f73d-4975-9444-e6efe1c5dba8" containerID="ff887759f80f5295b52f810d14bd483d3d05170a2a66591a4cc3d8544afd5eb8" exitCode=0 Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.948058 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k8hjh" event={"ID":"19488820-f73d-4975-9444-e6efe1c5dba8","Type":"ContainerDied","Data":"ff887759f80f5295b52f810d14bd483d3d05170a2a66591a4cc3d8544afd5eb8"} Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.948090 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k8hjh" Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.948114 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k8hjh" event={"ID":"19488820-f73d-4975-9444-e6efe1c5dba8","Type":"ContainerDied","Data":"e29a558cc02d1f830786e0b79ee76606023e401350bf4d9008813a7a8140ec4a"} Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.948141 4739 scope.go:117] "RemoveContainer" containerID="ff887759f80f5295b52f810d14bd483d3d05170a2a66591a4cc3d8544afd5eb8" Dec 01 16:54:20 crc kubenswrapper[4739]: I1201 16:54:20.979301 4739 scope.go:117] "RemoveContainer" containerID="09956881a68a2123539f0fcd20c25975c5074dfd44f181b1c88346ce90696b49" Dec 01 16:54:21 crc kubenswrapper[4739]: I1201 16:54:21.012949 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k8hjh"] Dec 01 16:54:21 crc kubenswrapper[4739]: I1201 16:54:21.022158 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k8hjh"] Dec 01 16:54:21 crc kubenswrapper[4739]: I1201 16:54:21.024966 4739 scope.go:117] "RemoveContainer" containerID="bbf2d1483f512131b96ad6bb670f2cda85aed21ca6b3c8d8b818536f5c2096ea" Dec 01 16:54:21 crc kubenswrapper[4739]: I1201 16:54:21.050213 4739 scope.go:117] "RemoveContainer" containerID="ff887759f80f5295b52f810d14bd483d3d05170a2a66591a4cc3d8544afd5eb8" Dec 01 16:54:21 crc kubenswrapper[4739]: E1201 16:54:21.050767 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff887759f80f5295b52f810d14bd483d3d05170a2a66591a4cc3d8544afd5eb8\": container with ID starting with ff887759f80f5295b52f810d14bd483d3d05170a2a66591a4cc3d8544afd5eb8 not found: ID does not exist" containerID="ff887759f80f5295b52f810d14bd483d3d05170a2a66591a4cc3d8544afd5eb8" Dec 01 16:54:21 crc kubenswrapper[4739]: I1201 16:54:21.050810 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff887759f80f5295b52f810d14bd483d3d05170a2a66591a4cc3d8544afd5eb8"} err="failed to get container status \"ff887759f80f5295b52f810d14bd483d3d05170a2a66591a4cc3d8544afd5eb8\": rpc error: code = NotFound desc = could not find container \"ff887759f80f5295b52f810d14bd483d3d05170a2a66591a4cc3d8544afd5eb8\": container with ID starting with ff887759f80f5295b52f810d14bd483d3d05170a2a66591a4cc3d8544afd5eb8 not found: ID does not exist" Dec 01 16:54:21 crc kubenswrapper[4739]: I1201 16:54:21.050840 4739 scope.go:117] "RemoveContainer" containerID="09956881a68a2123539f0fcd20c25975c5074dfd44f181b1c88346ce90696b49" Dec 01 16:54:21 crc kubenswrapper[4739]: E1201 16:54:21.051542 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09956881a68a2123539f0fcd20c25975c5074dfd44f181b1c88346ce90696b49\": container with ID starting with 09956881a68a2123539f0fcd20c25975c5074dfd44f181b1c88346ce90696b49 not found: ID does not exist" containerID="09956881a68a2123539f0fcd20c25975c5074dfd44f181b1c88346ce90696b49" Dec 01 16:54:21 crc kubenswrapper[4739]: I1201 16:54:21.051597 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09956881a68a2123539f0fcd20c25975c5074dfd44f181b1c88346ce90696b49"} err="failed to get container status \"09956881a68a2123539f0fcd20c25975c5074dfd44f181b1c88346ce90696b49\": rpc error: code = NotFound desc = could not find container \"09956881a68a2123539f0fcd20c25975c5074dfd44f181b1c88346ce90696b49\": container with ID starting with 09956881a68a2123539f0fcd20c25975c5074dfd44f181b1c88346ce90696b49 not found: ID does not exist" Dec 01 16:54:21 crc kubenswrapper[4739]: I1201 16:54:21.051628 4739 scope.go:117] "RemoveContainer" containerID="bbf2d1483f512131b96ad6bb670f2cda85aed21ca6b3c8d8b818536f5c2096ea" Dec 01 16:54:21 crc kubenswrapper[4739]: E1201 16:54:21.052242 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbf2d1483f512131b96ad6bb670f2cda85aed21ca6b3c8d8b818536f5c2096ea\": container with ID starting with bbf2d1483f512131b96ad6bb670f2cda85aed21ca6b3c8d8b818536f5c2096ea not found: ID does not exist" containerID="bbf2d1483f512131b96ad6bb670f2cda85aed21ca6b3c8d8b818536f5c2096ea" Dec 01 16:54:21 crc kubenswrapper[4739]: I1201 16:54:21.052341 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbf2d1483f512131b96ad6bb670f2cda85aed21ca6b3c8d8b818536f5c2096ea"} err="failed to get container status \"bbf2d1483f512131b96ad6bb670f2cda85aed21ca6b3c8d8b818536f5c2096ea\": rpc error: code = NotFound desc = could not find container \"bbf2d1483f512131b96ad6bb670f2cda85aed21ca6b3c8d8b818536f5c2096ea\": container with ID starting with bbf2d1483f512131b96ad6bb670f2cda85aed21ca6b3c8d8b818536f5c2096ea not found: ID does not exist" Dec 01 16:54:22 crc kubenswrapper[4739]: I1201 16:54:22.494942 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19488820-f73d-4975-9444-e6efe1c5dba8" path="/var/lib/kubelet/pods/19488820-f73d-4975-9444-e6efe1c5dba8/volumes" Dec 01 16:54:25 crc kubenswrapper[4739]: I1201 16:54:25.886296 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-856bfcfff6-5wtcc_b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff/barbican-api/0.log" Dec 01 16:54:26 crc kubenswrapper[4739]: I1201 16:54:26.079943 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-ff455997d-9lg5c_2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd/barbican-keystone-listener/0.log" Dec 01 16:54:26 crc kubenswrapper[4739]: I1201 16:54:26.137836 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-856bfcfff6-5wtcc_b34d197e-14ff-4c2e-b1e2-a6fde0dde4ff/barbican-api-log/0.log" Dec 01 16:54:26 crc kubenswrapper[4739]: I1201 16:54:26.377889 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6bbd6b9f5c-wjbz6_02f9d262-1e44-42e4-a413-b85bb5f25241/barbican-worker/0.log" Dec 01 16:54:26 crc kubenswrapper[4739]: I1201 16:54:26.389110 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6bbd6b9f5c-wjbz6_02f9d262-1e44-42e4-a413-b85bb5f25241/barbican-worker-log/0.log" Dec 01 16:54:26 crc kubenswrapper[4739]: I1201 16:54:26.435121 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-ff455997d-9lg5c_2b2e43dc-f874-4da1-aa0e-ed2d1c07aafd/barbican-keystone-listener-log/0.log" Dec 01 16:54:26 crc kubenswrapper[4739]: I1201 16:54:26.604706 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-97b2w_df49ac71-260c-49a1-ac48-ab9e863e08be/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 16:54:26 crc kubenswrapper[4739]: I1201 16:54:26.651800 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7bacd19a-d9e6-4894-bc04-c11b7c5f3c85/ceilometer-central-agent/0.log" Dec 01 16:54:26 crc kubenswrapper[4739]: I1201 16:54:26.795847 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7bacd19a-d9e6-4894-bc04-c11b7c5f3c85/proxy-httpd/0.log" Dec 01 16:54:26 crc kubenswrapper[4739]: I1201 16:54:26.807497 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7bacd19a-d9e6-4894-bc04-c11b7c5f3c85/ceilometer-notification-agent/0.log" Dec 01 16:54:26 crc kubenswrapper[4739]: I1201 16:54:26.813057 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_7bacd19a-d9e6-4894-bc04-c11b7c5f3c85/sg-core/0.log" Dec 01 16:54:26 crc kubenswrapper[4739]: I1201 16:54:26.992861 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jbqn7_af3127e6-d90c-4ac8-9e40-d8710d4f4bca/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 16:54:27 crc kubenswrapper[4739]: I1201 16:54:27.020830 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-k2llj_7c8c89e6-a763-42d7-8e47-84d172df5822/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 16:54:27 crc kubenswrapper[4739]: I1201 16:54:27.654669 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0630ef7a-359c-49a5-9906-b7c9e01936f7/probe/0.log" Dec 01 16:54:27 crc kubenswrapper[4739]: I1201 16:54:27.805903 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_32d670bc-bba0-45ed-80c9-f4258d97452c/cinder-api/0.log" Dec 01 16:54:27 crc kubenswrapper[4739]: I1201 16:54:27.987075 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_32d670bc-bba0-45ed-80c9-f4258d97452c/cinder-api-log/0.log" Dec 01 16:54:28 crc kubenswrapper[4739]: I1201 16:54:28.070202 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2/cinder-scheduler/0.log" Dec 01 16:54:28 crc kubenswrapper[4739]: I1201 16:54:28.131096 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4e7d4195-d816-4e99-bf1c-b4daa1b5e2c2/probe/0.log" Dec 01 16:54:28 crc kubenswrapper[4739]: I1201 16:54:28.348538 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_694f34a3-1680-4a0f-9320-9c0d83d957a7/probe/0.log" Dec 01 16:54:28 crc kubenswrapper[4739]: I1201 16:54:28.600384 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-v7d2m_0c99055a-3919-465d-8584-a82a814a4be7/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 16:54:28 crc kubenswrapper[4739]: I1201 16:54:28.841879 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-vcwgs_7881548e-3e40-4ffc-85b5-d64050ea26c6/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 16:54:29 crc kubenswrapper[4739]: I1201 16:54:29.046301 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-pvgg8_f27b159a-6738-43d0-8ba1-041cc46170dc/init/0.log" Dec 01 16:54:29 crc kubenswrapper[4739]: I1201 16:54:29.730205 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-pvgg8_f27b159a-6738-43d0-8ba1-041cc46170dc/init/0.log" Dec 01 16:54:29 crc kubenswrapper[4739]: I1201 16:54:29.873026 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76b5fdb995-pvgg8_f27b159a-6738-43d0-8ba1-041cc46170dc/dnsmasq-dns/0.log" Dec 01 16:54:29 crc kubenswrapper[4739]: I1201 16:54:29.978228 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_47d0ae7f-f45b-40a0-870b-f1d7c6252e45/glance-httpd/0.log" Dec 01 16:54:30 crc kubenswrapper[4739]: I1201 16:54:30.047976 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_47d0ae7f-f45b-40a0-870b-f1d7c6252e45/glance-log/0.log" Dec 01 16:54:30 crc kubenswrapper[4739]: I1201 16:54:30.250194 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_201539c2-475c-4512-912b-729e12b1ba88/glance-httpd/0.log" Dec 01 16:54:30 crc kubenswrapper[4739]: I1201 16:54:30.304290 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_201539c2-475c-4512-912b-729e12b1ba88/glance-log/0.log" Dec 01 16:54:30 crc kubenswrapper[4739]: I1201 16:54:30.642775 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-698d484b74-5hhp4_77ed99e7-001f-4dd6-9e00-5b83b7ab180f/horizon/0.log" Dec 01 16:54:30 crc kubenswrapper[4739]: I1201 16:54:30.868818 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-jpnft_d04ca357-3e15-4fbc-bfe1-299c3cbc2f0c/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 16:54:30 crc kubenswrapper[4739]: I1201 16:54:30.880568 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-698d484b74-5hhp4_77ed99e7-001f-4dd6-9e00-5b83b7ab180f/horizon-log/0.log" Dec 01 16:54:31 crc kubenswrapper[4739]: I1201 16:54:31.167342 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0630ef7a-359c-49a5-9906-b7c9e01936f7/cinder-backup/0.log" Dec 01 16:54:31 crc kubenswrapper[4739]: I1201 16:54:31.664640 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-k9trm_d28dec39-8ad1-4891-82bd-ce6080d608b8/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 16:54:31 crc kubenswrapper[4739]: I1201 16:54:31.939859 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29410081-vqlxx_8a3b97a2-1cd2-49ae-85b4-bc5e7fca2008/keystone-cron/0.log" Dec 01 16:54:32 crc kubenswrapper[4739]: I1201 16:54:32.117917 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_85e3c95a-b796-44da-93fe-e958ec7ae3c1/kube-state-metrics/0.log" Dec 01 16:54:32 crc kubenswrapper[4739]: I1201 16:54:32.363670 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-x85w7_cdf5f5e1-60c9-4db6-9e32-1b0ea68a6f7c/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 16:54:32 crc kubenswrapper[4739]: I1201 16:54:32.626294 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_ff7067bb-b693-4d14-9a83-af47d37b421a/manila-api/0.log" Dec 01 16:54:32 crc kubenswrapper[4739]: I1201 16:54:32.628972 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_ff7067bb-b693-4d14-9a83-af47d37b421a/manila-api-log/0.log" Dec 01 16:54:32 crc kubenswrapper[4739]: I1201 16:54:32.686863 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5f489c7bf6-cm54s_a244faee-6798-4c21-8041-6f5811a88eac/keystone-api/0.log" Dec 01 16:54:32 crc kubenswrapper[4739]: I1201 16:54:32.921347 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_be94c9f8-1f60-4cd9-b6f7-25459823386c/probe/0.log" Dec 01 16:54:32 crc kubenswrapper[4739]: I1201 16:54:32.951750 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_be94c9f8-1f60-4cd9-b6f7-25459823386c/manila-scheduler/0.log" Dec 01 16:54:32 crc kubenswrapper[4739]: I1201 16:54:32.974157 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_cf03c1ab-74fb-412e-9625-3993db3e51df/manila-share/0.log" Dec 01 16:54:33 crc kubenswrapper[4739]: I1201 16:54:33.149689 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_cf03c1ab-74fb-412e-9625-3993db3e51df/probe/0.log" Dec 01 16:54:33 crc kubenswrapper[4739]: I1201 16:54:33.720021 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d98bc475f-59p2r_cb777746-9ae6-4eb5-8459-aa82ad2eafb5/neutron-httpd/0.log" Dec 01 16:54:33 crc kubenswrapper[4739]: I1201 16:54:33.787373 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-ffjp2_529cfd75-6eac-42f1-ad43-f9c0ec3b7ff4/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 16:54:33 crc kubenswrapper[4739]: I1201 16:54:33.794641 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d98bc475f-59p2r_cb777746-9ae6-4eb5-8459-aa82ad2eafb5/neutron-api/0.log" Dec 01 16:54:34 crc kubenswrapper[4739]: I1201 16:54:34.834198 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_50a7f5a5-c069-4e82-97e0-ecccdf443c15/nova-cell0-conductor-conductor/0.log" Dec 01 16:54:34 crc kubenswrapper[4739]: I1201 16:54:34.931383 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_31d2c216-973d-47da-ae9b-2cdaa7f375ee/nova-api-log/0.log" Dec 01 16:54:35 crc kubenswrapper[4739]: I1201 16:54:35.257635 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_407b0796-c7a3-4ae4-903d-c8173b6cc378/nova-cell1-conductor-conductor/0.log" Dec 01 16:54:35 crc kubenswrapper[4739]: I1201 16:54:35.389784 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_31d2c216-973d-47da-ae9b-2cdaa7f375ee/nova-api-api/0.log" Dec 01 16:54:35 crc kubenswrapper[4739]: I1201 16:54:35.586117 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_ab4c4993-67b7-46f2-8aeb-0173ffc9d1a0/nova-cell1-novncproxy-novncproxy/0.log" Dec 01 16:54:35 crc kubenswrapper[4739]: I1201 16:54:35.697175 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-7hvg4_ad6c9132-a47c-486c-bb34-17d977bef91b/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 16:54:35 crc kubenswrapper[4739]: I1201 16:54:35.871939 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9/nova-metadata-log/0.log" Dec 01 16:54:36 crc kubenswrapper[4739]: I1201 16:54:36.357672 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_25515423-047d-445c-bae8-f54adf3ce705/nova-scheduler-scheduler/0.log" Dec 01 16:54:36 crc kubenswrapper[4739]: I1201 16:54:36.544625 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a5323c11-dc67-47e7-83be-4b8a52f4ebef/mysql-bootstrap/0.log" Dec 01 16:54:36 crc kubenswrapper[4739]: I1201 16:54:36.799379 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a5323c11-dc67-47e7-83be-4b8a52f4ebef/mysql-bootstrap/0.log" Dec 01 16:54:36 crc kubenswrapper[4739]: I1201 16:54:36.810503 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a5323c11-dc67-47e7-83be-4b8a52f4ebef/galera/0.log" Dec 01 16:54:37 crc kubenswrapper[4739]: I1201 16:54:37.008654 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8227dcb8-9b5f-4348-bba7-1560fa84861a/mysql-bootstrap/0.log" Dec 01 16:54:37 crc kubenswrapper[4739]: I1201 16:54:37.189646 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8227dcb8-9b5f-4348-bba7-1560fa84861a/mysql-bootstrap/0.log" Dec 01 16:54:37 crc kubenswrapper[4739]: I1201 16:54:37.287503 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8227dcb8-9b5f-4348-bba7-1560fa84861a/galera/0.log" Dec 01 16:54:37 crc kubenswrapper[4739]: I1201 16:54:37.481846 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_ef1ce8cc-853d-415c-85be-39fc10bec74d/openstackclient/0.log" Dec 01 16:54:37 crc kubenswrapper[4739]: I1201 16:54:37.681927 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4ltrp_10ee4b3e-c925-4d5d-9453-7b3c9fd2ed15/ovn-controller/0.log" Dec 01 16:54:37 crc kubenswrapper[4739]: I1201 16:54:37.884779 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kznqt_1ea74040-7f82-45a2-a76d-39aca5a88282/openstack-network-exporter/0.log" Dec 01 16:54:38 crc kubenswrapper[4739]: I1201 16:54:38.163486 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pr8vl_0bed2a41-e961-4361-9316-2f29b6c0b0b5/ovsdb-server-init/0.log" Dec 01 16:54:38 crc kubenswrapper[4739]: I1201 16:54:38.285083 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0e21a1fb-5536-4eb8-8dd5-eeea64d1f6f9/nova-metadata-metadata/0.log" Dec 01 16:54:38 crc kubenswrapper[4739]: I1201 16:54:38.334555 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_694f34a3-1680-4a0f-9320-9c0d83d957a7/cinder-volume/0.log" Dec 01 16:54:38 crc kubenswrapper[4739]: I1201 16:54:38.382781 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pr8vl_0bed2a41-e961-4361-9316-2f29b6c0b0b5/ovsdb-server-init/0.log" Dec 01 16:54:38 crc kubenswrapper[4739]: I1201 16:54:38.385095 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pr8vl_0bed2a41-e961-4361-9316-2f29b6c0b0b5/ovs-vswitchd/0.log" Dec 01 16:54:38 crc kubenswrapper[4739]: I1201 16:54:38.486494 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pr8vl_0bed2a41-e961-4361-9316-2f29b6c0b0b5/ovsdb-server/0.log" Dec 01 16:54:38 crc kubenswrapper[4739]: I1201 16:54:38.622572 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-5sh69_58d30160-c6ad-4f46-bc27-f7b28a07d97a/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 16:54:38 crc kubenswrapper[4739]: I1201 16:54:38.711954 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d/openstack-network-exporter/0.log" Dec 01 16:54:38 crc kubenswrapper[4739]: I1201 16:54:38.817833 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_cd8d8f3a-31a9-4709-a0d4-b7edd8c9937d/ovn-northd/0.log" Dec 01 16:54:38 crc kubenswrapper[4739]: I1201 16:54:38.880260 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_17dd4503-6f0a-48f1-b661-d78154c9a05d/openstack-network-exporter/0.log" Dec 01 16:54:38 crc kubenswrapper[4739]: I1201 16:54:38.902615 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_17dd4503-6f0a-48f1-b661-d78154c9a05d/ovsdbserver-nb/0.log" Dec 01 16:54:39 crc kubenswrapper[4739]: I1201 16:54:39.127892 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_723e3e8b-e65e-46d6-9cd5-11b0d176e8c7/openstack-network-exporter/0.log" Dec 01 16:54:39 crc kubenswrapper[4739]: I1201 16:54:39.130394 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_723e3e8b-e65e-46d6-9cd5-11b0d176e8c7/ovsdbserver-sb/0.log" Dec 01 16:54:39 crc kubenswrapper[4739]: I1201 16:54:39.456900 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-54db66b7f6-6plgv_cc9d7a70-2c80-4443-bd55-be94d972c336/placement-api/0.log" Dec 01 16:54:39 crc kubenswrapper[4739]: I1201 16:54:39.459201 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c4dbaaa4-ac29-4909-9200-6db94e30f4e5/setup-container/0.log" Dec 01 16:54:39 crc kubenswrapper[4739]: I1201 16:54:39.482900 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-54db66b7f6-6plgv_cc9d7a70-2c80-4443-bd55-be94d972c336/placement-log/0.log" Dec 01 16:54:39 crc kubenswrapper[4739]: I1201 16:54:39.587106 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c4dbaaa4-ac29-4909-9200-6db94e30f4e5/setup-container/0.log" Dec 01 16:54:39 crc kubenswrapper[4739]: I1201 16:54:39.689867 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c4dbaaa4-ac29-4909-9200-6db94e30f4e5/rabbitmq/0.log" Dec 01 16:54:39 crc kubenswrapper[4739]: I1201 16:54:39.737065 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1ad80536-1ee5-4fcd-9de2-16b7683f1870/setup-container/0.log" Dec 01 16:54:40 crc kubenswrapper[4739]: I1201 16:54:40.572096 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1ad80536-1ee5-4fcd-9de2-16b7683f1870/setup-container/0.log" Dec 01 16:54:40 crc kubenswrapper[4739]: I1201 16:54:40.728037 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1ad80536-1ee5-4fcd-9de2-16b7683f1870/rabbitmq/0.log" Dec 01 16:54:40 crc kubenswrapper[4739]: I1201 16:54:40.749465 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-56zf8_f111abd9-3325-4578-8429-f107f9ba3185/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 16:54:40 crc kubenswrapper[4739]: I1201 16:54:40.997122 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-x8s8n_07317009-6689-44fd-8d13-1b022cc3e4d1/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 16:54:41 crc kubenswrapper[4739]: I1201 16:54:41.039114 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-wstvs_d73d5b54-3d1a-444e-acf0-4f2c7bfedf3a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 16:54:41 crc kubenswrapper[4739]: I1201 16:54:41.248999 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-vq2hk_b45dd823-9e52-42f9-8bcf-5b27757f46e2/ssh-known-hosts-edpm-deployment/0.log" Dec 01 16:54:41 crc kubenswrapper[4739]: I1201 16:54:41.342210 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_d7438aa0-1e45-4112-9af4-a584825d29df/tempest-tests-tempest-tests-runner/0.log" Dec 01 16:54:41 crc kubenswrapper[4739]: I1201 16:54:41.468647 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_b78ab401-a093-4c5d-aaa2-4c95534c3071/test-operator-logs-container/0.log" Dec 01 16:54:41 crc kubenswrapper[4739]: I1201 16:54:41.636377 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-kdcxk_ddf28305-5484-4d69-be7c-04fe79db3741/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 16:54:52 crc kubenswrapper[4739]: I1201 16:54:52.625288 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_7301c45d-f6e6-4d47-b75d-d381033edb1d/memcached/0.log" Dec 01 16:55:09 crc kubenswrapper[4739]: I1201 16:55:09.953201 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp_500c937c-6baf-47e7-93ec-fbff3d5b6d37/util/0.log" Dec 01 16:55:10 crc kubenswrapper[4739]: I1201 16:55:10.131048 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp_500c937c-6baf-47e7-93ec-fbff3d5b6d37/util/0.log" Dec 01 16:55:10 crc kubenswrapper[4739]: I1201 16:55:10.175400 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp_500c937c-6baf-47e7-93ec-fbff3d5b6d37/pull/0.log" Dec 01 16:55:10 crc kubenswrapper[4739]: I1201 16:55:10.209261 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp_500c937c-6baf-47e7-93ec-fbff3d5b6d37/pull/0.log" Dec 01 16:55:10 crc kubenswrapper[4739]: I1201 16:55:10.754480 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp_500c937c-6baf-47e7-93ec-fbff3d5b6d37/pull/0.log" Dec 01 16:55:10 crc kubenswrapper[4739]: I1201 16:55:10.787916 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp_500c937c-6baf-47e7-93ec-fbff3d5b6d37/extract/0.log" Dec 01 16:55:10 crc kubenswrapper[4739]: I1201 16:55:10.792453 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b44dc05e0ad3c1fe3f6722b8f09ca0621070b9c4a4ad26d5697a0486b4mr9fp_500c937c-6baf-47e7-93ec-fbff3d5b6d37/util/0.log" Dec 01 16:55:11 crc kubenswrapper[4739]: I1201 16:55:11.157499 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-nz6dz_97f978b1-c0ba-4fa7-8a8f-91de3ab80bc5/kube-rbac-proxy/0.log" Dec 01 16:55:11 crc kubenswrapper[4739]: I1201 16:55:11.217194 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-nz6dz_97f978b1-c0ba-4fa7-8a8f-91de3ab80bc5/manager/0.log" Dec 01 16:55:11 crc kubenswrapper[4739]: I1201 16:55:11.227820 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d986d85c9-2s7sq_28db1029-7abe-4d58-afbd-0022422c286c/kube-rbac-proxy/0.log" Dec 01 16:55:11 crc kubenswrapper[4739]: I1201 16:55:11.453364 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d986d85c9-2s7sq_28db1029-7abe-4d58-afbd-0022422c286c/manager/0.log" Dec 01 16:55:11 crc kubenswrapper[4739]: I1201 16:55:11.479187 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-th7k2_44a9534e-febc-4a41-9ee0-af96baa95a78/kube-rbac-proxy/0.log" Dec 01 16:55:11 crc kubenswrapper[4739]: I1201 16:55:11.495194 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-th7k2_44a9534e-febc-4a41-9ee0-af96baa95a78/manager/0.log" Dec 01 16:55:11 crc kubenswrapper[4739]: I1201 16:55:11.698003 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-2m26g_8525c2bf-aedc-4816-8ea5-4ebdda160f3d/kube-rbac-proxy/0.log" Dec 01 16:55:11 crc kubenswrapper[4739]: I1201 16:55:11.773372 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-2m26g_8525c2bf-aedc-4816-8ea5-4ebdda160f3d/manager/0.log" Dec 01 16:55:11 crc kubenswrapper[4739]: I1201 16:55:11.912460 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-tmntr_3422b21c-7fef-4c3a-8990-85fa0d40b027/kube-rbac-proxy/0.log" Dec 01 16:55:11 crc kubenswrapper[4739]: I1201 16:55:11.964005 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-tmntr_3422b21c-7fef-4c3a-8990-85fa0d40b027/manager/0.log" Dec 01 16:55:11 crc kubenswrapper[4739]: I1201 16:55:11.991173 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-f2t8s_895e540c-7e60-45b1-8275-36588bb8e788/kube-rbac-proxy/0.log" Dec 01 16:55:12 crc kubenswrapper[4739]: I1201 16:55:12.118288 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-f2t8s_895e540c-7e60-45b1-8275-36588bb8e788/manager/0.log" Dec 01 16:55:12 crc kubenswrapper[4739]: I1201 16:55:12.249333 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-gpkcn_54dcd8b3-40f6-4487-bc4f-810c6aebb6fe/kube-rbac-proxy/0.log" Dec 01 16:55:12 crc kubenswrapper[4739]: I1201 16:55:12.371503 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-gpkcn_54dcd8b3-40f6-4487-bc4f-810c6aebb6fe/manager/0.log" Dec 01 16:55:12 crc kubenswrapper[4739]: I1201 16:55:12.470207 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-pqqzm_ebe9e6cb-7cc7-4a43-ba59-ff0b8209d314/kube-rbac-proxy/0.log" Dec 01 16:55:12 crc kubenswrapper[4739]: I1201 16:55:12.484032 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-pqqzm_ebe9e6cb-7cc7-4a43-ba59-ff0b8209d314/manager/0.log" Dec 01 16:55:12 crc kubenswrapper[4739]: I1201 16:55:12.673578 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-758wd_8f8c53db-885f-4548-bcc0-02ce36cd6543/kube-rbac-proxy/0.log" Dec 01 16:55:12 crc kubenswrapper[4739]: I1201 16:55:12.704104 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-758wd_8f8c53db-885f-4548-bcc0-02ce36cd6543/manager/0.log" Dec 01 16:55:12 crc kubenswrapper[4739]: I1201 16:55:12.900120 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-d2dxq_7a27f279-7c57-439c-88f2-67914f2e7b9e/kube-rbac-proxy/0.log" Dec 01 16:55:12 crc kubenswrapper[4739]: I1201 16:55:12.942574 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-d2dxq_7a27f279-7c57-439c-88f2-67914f2e7b9e/manager/0.log" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.035449 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9f2gf"] Dec 01 16:55:13 crc kubenswrapper[4739]: E1201 16:55:13.035926 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19488820-f73d-4975-9444-e6efe1c5dba8" containerName="extract-content" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.035946 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="19488820-f73d-4975-9444-e6efe1c5dba8" containerName="extract-content" Dec 01 16:55:13 crc kubenswrapper[4739]: E1201 16:55:13.035975 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19488820-f73d-4975-9444-e6efe1c5dba8" containerName="extract-utilities" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.035983 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="19488820-f73d-4975-9444-e6efe1c5dba8" containerName="extract-utilities" Dec 01 16:55:13 crc kubenswrapper[4739]: E1201 16:55:13.035997 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19488820-f73d-4975-9444-e6efe1c5dba8" containerName="registry-server" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.036004 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="19488820-f73d-4975-9444-e6efe1c5dba8" containerName="registry-server" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.036192 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="19488820-f73d-4975-9444-e6efe1c5dba8" containerName="registry-server" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.037767 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.047473 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9f2gf"] Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.099213 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/632e0bd0-9997-470e-81b2-88475ce103f0-utilities\") pod \"redhat-marketplace-9f2gf\" (UID: \"632e0bd0-9997-470e-81b2-88475ce103f0\") " pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.099295 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpxgp\" (UniqueName: \"kubernetes.io/projected/632e0bd0-9997-470e-81b2-88475ce103f0-kube-api-access-vpxgp\") pod \"redhat-marketplace-9f2gf\" (UID: \"632e0bd0-9997-470e-81b2-88475ce103f0\") " pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.099360 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/632e0bd0-9997-470e-81b2-88475ce103f0-catalog-content\") pod \"redhat-marketplace-9f2gf\" (UID: \"632e0bd0-9997-470e-81b2-88475ce103f0\") " pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.109327 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-g7bc7_644da74f-c26b-4fae-a81f-4091a1e95b87/kube-rbac-proxy/0.log" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.200748 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/632e0bd0-9997-470e-81b2-88475ce103f0-utilities\") pod \"redhat-marketplace-9f2gf\" (UID: \"632e0bd0-9997-470e-81b2-88475ce103f0\") " pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.200804 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpxgp\" (UniqueName: \"kubernetes.io/projected/632e0bd0-9997-470e-81b2-88475ce103f0-kube-api-access-vpxgp\") pod \"redhat-marketplace-9f2gf\" (UID: \"632e0bd0-9997-470e-81b2-88475ce103f0\") " pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.200845 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/632e0bd0-9997-470e-81b2-88475ce103f0-catalog-content\") pod \"redhat-marketplace-9f2gf\" (UID: \"632e0bd0-9997-470e-81b2-88475ce103f0\") " pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.201387 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/632e0bd0-9997-470e-81b2-88475ce103f0-catalog-content\") pod \"redhat-marketplace-9f2gf\" (UID: \"632e0bd0-9997-470e-81b2-88475ce103f0\") " pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.201679 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/632e0bd0-9997-470e-81b2-88475ce103f0-utilities\") pod \"redhat-marketplace-9f2gf\" (UID: \"632e0bd0-9997-470e-81b2-88475ce103f0\") " pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.208847 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-g7bc7_644da74f-c26b-4fae-a81f-4091a1e95b87/manager/0.log" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.241290 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpxgp\" (UniqueName: \"kubernetes.io/projected/632e0bd0-9997-470e-81b2-88475ce103f0-kube-api-access-vpxgp\") pod \"redhat-marketplace-9f2gf\" (UID: \"632e0bd0-9997-470e-81b2-88475ce103f0\") " pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.312502 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-jrc6n_5d6c8b0a-e856-4794-a3ce-f72e1f6d0c91/kube-rbac-proxy/0.log" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.363655 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.393714 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-jrc6n_5d6c8b0a-e856-4794-a3ce-f72e1f6d0c91/manager/0.log" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.685999 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-8bsmt_974d1918-6e6a-4660-898c-b8d8db439f46/manager/0.log" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.694048 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-8bsmt_974d1918-6e6a-4660-898c-b8d8db439f46/kube-rbac-proxy/0.log" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.824749 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-kkchm_a4706184-6b4f-4416-b9d6-430087d5577c/kube-rbac-proxy/0.log" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.861065 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-kkchm_a4706184-6b4f-4416-b9d6-430087d5577c/manager/0.log" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.931271 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9f2gf"] Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.982697 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh_81c50b9d-cef2-40b0-bb96-0e44676eaf93/kube-rbac-proxy/0.log" Dec 01 16:55:13 crc kubenswrapper[4739]: I1201 16:55:13.995279 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4blnvh_81c50b9d-cef2-40b0-bb96-0e44676eaf93/manager/0.log" Dec 01 16:55:14 crc kubenswrapper[4739]: I1201 16:55:14.404834 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-sqnf9_38434f68-049c-4b26-ad64-c6da629a6df2/registry-server/0.log" Dec 01 16:55:14 crc kubenswrapper[4739]: I1201 16:55:14.429870 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-b8d979694-j8gm2_ee802613-91de-4f5b-b0c8-7966498b954c/operator/0.log" Dec 01 16:55:14 crc kubenswrapper[4739]: I1201 16:55:14.472478 4739 generic.go:334] "Generic (PLEG): container finished" podID="632e0bd0-9997-470e-81b2-88475ce103f0" containerID="d96801347fdb8ec470810df69d89b6df74bb78c8a5017919f5e8a0b3e243de7f" exitCode=0 Dec 01 16:55:14 crc kubenswrapper[4739]: I1201 16:55:14.472513 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9f2gf" event={"ID":"632e0bd0-9997-470e-81b2-88475ce103f0","Type":"ContainerDied","Data":"d96801347fdb8ec470810df69d89b6df74bb78c8a5017919f5e8a0b3e243de7f"} Dec 01 16:55:14 crc kubenswrapper[4739]: I1201 16:55:14.472551 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9f2gf" event={"ID":"632e0bd0-9997-470e-81b2-88475ce103f0","Type":"ContainerStarted","Data":"1cf66bab2dbe759665e194f2ffb082465171fd65b343512fd5b4c0dfa9ea1cf3"} Dec 01 16:55:14 crc kubenswrapper[4739]: I1201 16:55:14.473571 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-4nwgp_903f2159-9ae0-4b78-8447-3b5bd4a420ef/kube-rbac-proxy/0.log" Dec 01 16:55:14 crc kubenswrapper[4739]: I1201 16:55:14.475030 4739 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:55:14 crc kubenswrapper[4739]: I1201 16:55:14.698446 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-4nwgp_903f2159-9ae0-4b78-8447-3b5bd4a420ef/manager/0.log" Dec 01 16:55:14 crc kubenswrapper[4739]: I1201 16:55:14.707180 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-zrhx2_c6ddb146-59dc-41c2-9278-466eda24296c/kube-rbac-proxy/0.log" Dec 01 16:55:14 crc kubenswrapper[4739]: I1201 16:55:14.800026 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-zrhx2_c6ddb146-59dc-41c2-9278-466eda24296c/manager/0.log" Dec 01 16:55:14 crc kubenswrapper[4739]: I1201 16:55:14.984903 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-qm8fw_47bf6f4a-97c7-4f1f-a00d-0c5151058f2c/operator/0.log" Dec 01 16:55:15 crc kubenswrapper[4739]: I1201 16:55:15.022932 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-xgpkt_7ff65da3-52d5-4e80-9f48-d23d3bc22ff6/kube-rbac-proxy/0.log" Dec 01 16:55:15 crc kubenswrapper[4739]: I1201 16:55:15.181324 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-xgpkt_7ff65da3-52d5-4e80-9f48-d23d3bc22ff6/manager/0.log" Dec 01 16:55:15 crc kubenswrapper[4739]: I1201 16:55:15.231287 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-fxvzt_f60761e0-1afe-46fa-90fc-5ed9f42f730d/kube-rbac-proxy/0.log" Dec 01 16:55:15 crc kubenswrapper[4739]: I1201 16:55:15.394572 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-fxvzt_f60761e0-1afe-46fa-90fc-5ed9f42f730d/manager/0.log" Dec 01 16:55:15 crc kubenswrapper[4739]: I1201 16:55:15.410872 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-4pwqm_bed989ab-cd94-41f1-91ea-8c1bdfc056af/kube-rbac-proxy/0.log" Dec 01 16:55:15 crc kubenswrapper[4739]: I1201 16:55:15.434763 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-59fc5775d4-chxkc_5f474df5-ced8-45ff-840a-c52898696bf7/manager/0.log" Dec 01 16:55:15 crc kubenswrapper[4739]: I1201 16:55:15.485135 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9f2gf" event={"ID":"632e0bd0-9997-470e-81b2-88475ce103f0","Type":"ContainerStarted","Data":"b241eb5cc2912523531f85051bfb8a45822701881f4404121466cf7cd09697d1"} Dec 01 16:55:15 crc kubenswrapper[4739]: I1201 16:55:15.487902 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-4pwqm_bed989ab-cd94-41f1-91ea-8c1bdfc056af/manager/0.log" Dec 01 16:55:15 crc kubenswrapper[4739]: I1201 16:55:15.637158 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-px5zx_84c3dbfb-b9ac-4b9d-b942-e454b50bf132/kube-rbac-proxy/0.log" Dec 01 16:55:15 crc kubenswrapper[4739]: I1201 16:55:15.638990 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-px5zx_84c3dbfb-b9ac-4b9d-b942-e454b50bf132/manager/0.log" Dec 01 16:55:16 crc kubenswrapper[4739]: I1201 16:55:16.494939 4739 generic.go:334] "Generic (PLEG): container finished" podID="632e0bd0-9997-470e-81b2-88475ce103f0" containerID="b241eb5cc2912523531f85051bfb8a45822701881f4404121466cf7cd09697d1" exitCode=0 Dec 01 16:55:16 crc kubenswrapper[4739]: I1201 16:55:16.495021 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9f2gf" event={"ID":"632e0bd0-9997-470e-81b2-88475ce103f0","Type":"ContainerDied","Data":"b241eb5cc2912523531f85051bfb8a45822701881f4404121466cf7cd09697d1"} Dec 01 16:55:17 crc kubenswrapper[4739]: I1201 16:55:17.505609 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9f2gf" event={"ID":"632e0bd0-9997-470e-81b2-88475ce103f0","Type":"ContainerStarted","Data":"1634a346c9fc7d2aece547cef373f74e710f0765ee07ddb4391c0b180fd6914f"} Dec 01 16:55:17 crc kubenswrapper[4739]: I1201 16:55:17.559889 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9f2gf" podStartSLOduration=2.027223074 podStartE2EDuration="4.559860684s" podCreationTimestamp="2025-12-01 16:55:13 +0000 UTC" firstStartedPulling="2025-12-01 16:55:14.474766339 +0000 UTC m=+4816.300512433" lastFinishedPulling="2025-12-01 16:55:17.007403959 +0000 UTC m=+4818.833150043" observedRunningTime="2025-12-01 16:55:17.526369799 +0000 UTC m=+4819.352115903" watchObservedRunningTime="2025-12-01 16:55:17.559860684 +0000 UTC m=+4819.385606778" Dec 01 16:55:23 crc kubenswrapper[4739]: I1201 16:55:23.364882 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:23 crc kubenswrapper[4739]: I1201 16:55:23.366594 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:23 crc kubenswrapper[4739]: I1201 16:55:23.417729 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:23 crc kubenswrapper[4739]: I1201 16:55:23.611599 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:23 crc kubenswrapper[4739]: I1201 16:55:23.674835 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9f2gf"] Dec 01 16:55:25 crc kubenswrapper[4739]: I1201 16:55:25.574134 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9f2gf" podUID="632e0bd0-9997-470e-81b2-88475ce103f0" containerName="registry-server" containerID="cri-o://1634a346c9fc7d2aece547cef373f74e710f0765ee07ddb4391c0b180fd6914f" gracePeriod=2 Dec 01 16:55:26 crc kubenswrapper[4739]: I1201 16:55:26.587049 4739 generic.go:334] "Generic (PLEG): container finished" podID="632e0bd0-9997-470e-81b2-88475ce103f0" containerID="1634a346c9fc7d2aece547cef373f74e710f0765ee07ddb4391c0b180fd6914f" exitCode=0 Dec 01 16:55:26 crc kubenswrapper[4739]: I1201 16:55:26.587222 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9f2gf" event={"ID":"632e0bd0-9997-470e-81b2-88475ce103f0","Type":"ContainerDied","Data":"1634a346c9fc7d2aece547cef373f74e710f0765ee07ddb4391c0b180fd6914f"} Dec 01 16:55:26 crc kubenswrapper[4739]: I1201 16:55:26.587505 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9f2gf" event={"ID":"632e0bd0-9997-470e-81b2-88475ce103f0","Type":"ContainerDied","Data":"1cf66bab2dbe759665e194f2ffb082465171fd65b343512fd5b4c0dfa9ea1cf3"} Dec 01 16:55:26 crc kubenswrapper[4739]: I1201 16:55:26.587521 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cf66bab2dbe759665e194f2ffb082465171fd65b343512fd5b4c0dfa9ea1cf3" Dec 01 16:55:26 crc kubenswrapper[4739]: I1201 16:55:26.592047 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:26 crc kubenswrapper[4739]: I1201 16:55:26.678656 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/632e0bd0-9997-470e-81b2-88475ce103f0-utilities\") pod \"632e0bd0-9997-470e-81b2-88475ce103f0\" (UID: \"632e0bd0-9997-470e-81b2-88475ce103f0\") " Dec 01 16:55:26 crc kubenswrapper[4739]: I1201 16:55:26.678852 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/632e0bd0-9997-470e-81b2-88475ce103f0-catalog-content\") pod \"632e0bd0-9997-470e-81b2-88475ce103f0\" (UID: \"632e0bd0-9997-470e-81b2-88475ce103f0\") " Dec 01 16:55:26 crc kubenswrapper[4739]: I1201 16:55:26.678876 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpxgp\" (UniqueName: \"kubernetes.io/projected/632e0bd0-9997-470e-81b2-88475ce103f0-kube-api-access-vpxgp\") pod \"632e0bd0-9997-470e-81b2-88475ce103f0\" (UID: \"632e0bd0-9997-470e-81b2-88475ce103f0\") " Dec 01 16:55:26 crc kubenswrapper[4739]: I1201 16:55:26.679714 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/632e0bd0-9997-470e-81b2-88475ce103f0-utilities" (OuterVolumeSpecName: "utilities") pod "632e0bd0-9997-470e-81b2-88475ce103f0" (UID: "632e0bd0-9997-470e-81b2-88475ce103f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:55:26 crc kubenswrapper[4739]: I1201 16:55:26.683474 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/632e0bd0-9997-470e-81b2-88475ce103f0-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:55:26 crc kubenswrapper[4739]: I1201 16:55:26.695538 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/632e0bd0-9997-470e-81b2-88475ce103f0-kube-api-access-vpxgp" (OuterVolumeSpecName: "kube-api-access-vpxgp") pod "632e0bd0-9997-470e-81b2-88475ce103f0" (UID: "632e0bd0-9997-470e-81b2-88475ce103f0"). InnerVolumeSpecName "kube-api-access-vpxgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:55:26 crc kubenswrapper[4739]: I1201 16:55:26.697650 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/632e0bd0-9997-470e-81b2-88475ce103f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "632e0bd0-9997-470e-81b2-88475ce103f0" (UID: "632e0bd0-9997-470e-81b2-88475ce103f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:55:26 crc kubenswrapper[4739]: I1201 16:55:26.795331 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/632e0bd0-9997-470e-81b2-88475ce103f0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:55:26 crc kubenswrapper[4739]: I1201 16:55:26.795377 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpxgp\" (UniqueName: \"kubernetes.io/projected/632e0bd0-9997-470e-81b2-88475ce103f0-kube-api-access-vpxgp\") on node \"crc\" DevicePath \"\"" Dec 01 16:55:27 crc kubenswrapper[4739]: I1201 16:55:27.594630 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9f2gf" Dec 01 16:55:27 crc kubenswrapper[4739]: I1201 16:55:27.634269 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9f2gf"] Dec 01 16:55:27 crc kubenswrapper[4739]: I1201 16:55:27.647504 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9f2gf"] Dec 01 16:55:28 crc kubenswrapper[4739]: I1201 16:55:28.490674 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="632e0bd0-9997-470e-81b2-88475ce103f0" path="/var/lib/kubelet/pods/632e0bd0-9997-470e-81b2-88475ce103f0/volumes" Dec 01 16:55:35 crc kubenswrapper[4739]: I1201 16:55:35.541125 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-2jcd2_c0af2f6a-1320-4a72-9bfc-684d0f666756/control-plane-machine-set-operator/0.log" Dec 01 16:55:35 crc kubenswrapper[4739]: I1201 16:55:35.736785 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j47hk_93126f7a-dfe3-4864-bc30-ac3629eb2bc2/kube-rbac-proxy/0.log" Dec 01 16:55:35 crc kubenswrapper[4739]: I1201 16:55:35.749106 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j47hk_93126f7a-dfe3-4864-bc30-ac3629eb2bc2/machine-api-operator/0.log" Dec 01 16:55:50 crc kubenswrapper[4739]: I1201 16:55:50.198595 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-d4qnq_a3382af1-0e5e-41cc-9446-8f0a66d741cb/cert-manager-controller/0.log" Dec 01 16:55:50 crc kubenswrapper[4739]: I1201 16:55:50.381241 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-qp8rs_04583995-14a7-4263-ac42-3628b4e5760c/cert-manager-cainjector/0.log" Dec 01 16:55:50 crc kubenswrapper[4739]: I1201 16:55:50.452177 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-fm6tr_56175826-6fcb-4142-8716-e4fb8664a8b8/cert-manager-webhook/0.log" Dec 01 16:56:03 crc kubenswrapper[4739]: I1201 16:56:03.319323 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-8fmbb_5889130c-3ff0-4d1d-b78c-8bce4ca14615/nmstate-console-plugin/0.log" Dec 01 16:56:03 crc kubenswrapper[4739]: I1201 16:56:03.472142 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-cc6rd_cdc17bb8-f21a-4de6-a6d7-87a69f18333e/nmstate-handler/0.log" Dec 01 16:56:03 crc kubenswrapper[4739]: I1201 16:56:03.526446 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-6fdfz_d9a2f3d0-ea8e-435c-8e82-d1662c4cdeac/kube-rbac-proxy/0.log" Dec 01 16:56:03 crc kubenswrapper[4739]: I1201 16:56:03.552918 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-6fdfz_d9a2f3d0-ea8e-435c-8e82-d1662c4cdeac/nmstate-metrics/0.log" Dec 01 16:56:04 crc kubenswrapper[4739]: I1201 16:56:04.250825 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-b9wt5_7edd99c1-023c-4505-b716-c87a7f5d3deb/nmstate-operator/0.log" Dec 01 16:56:04 crc kubenswrapper[4739]: I1201 16:56:04.331870 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-j2vz5_aeaf3a4d-e661-4ed4-b54d-96b39645d838/nmstate-webhook/0.log" Dec 01 16:56:19 crc kubenswrapper[4739]: I1201 16:56:19.532706 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-t95pz_b839c085-2a05-4b9a-940d-341fcaa4df18/kube-rbac-proxy/0.log" Dec 01 16:56:19 crc kubenswrapper[4739]: I1201 16:56:19.562103 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-t95pz_b839c085-2a05-4b9a-940d-341fcaa4df18/controller/0.log" Dec 01 16:56:19 crc kubenswrapper[4739]: I1201 16:56:19.838849 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/cp-frr-files/0.log" Dec 01 16:56:20 crc kubenswrapper[4739]: I1201 16:56:20.076308 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/cp-frr-files/0.log" Dec 01 16:56:20 crc kubenswrapper[4739]: I1201 16:56:20.094669 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/cp-metrics/0.log" Dec 01 16:56:20 crc kubenswrapper[4739]: I1201 16:56:20.146870 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/cp-reloader/0.log" Dec 01 16:56:20 crc kubenswrapper[4739]: I1201 16:56:20.151982 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/cp-reloader/0.log" Dec 01 16:56:20 crc kubenswrapper[4739]: I1201 16:56:20.380014 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/cp-reloader/0.log" Dec 01 16:56:20 crc kubenswrapper[4739]: I1201 16:56:20.405069 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/cp-frr-files/0.log" Dec 01 16:56:20 crc kubenswrapper[4739]: I1201 16:56:20.455286 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/cp-metrics/0.log" Dec 01 16:56:20 crc kubenswrapper[4739]: I1201 16:56:20.464854 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/cp-metrics/0.log" Dec 01 16:56:20 crc kubenswrapper[4739]: I1201 16:56:20.685170 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/cp-frr-files/0.log" Dec 01 16:56:20 crc kubenswrapper[4739]: I1201 16:56:20.736134 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/cp-reloader/0.log" Dec 01 16:56:20 crc kubenswrapper[4739]: I1201 16:56:20.748721 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/controller/0.log" Dec 01 16:56:20 crc kubenswrapper[4739]: I1201 16:56:20.766934 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/cp-metrics/0.log" Dec 01 16:56:20 crc kubenswrapper[4739]: I1201 16:56:20.977593 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/frr-metrics/0.log" Dec 01 16:56:21 crc kubenswrapper[4739]: I1201 16:56:21.021951 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/kube-rbac-proxy-frr/0.log" Dec 01 16:56:21 crc kubenswrapper[4739]: I1201 16:56:21.063338 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/kube-rbac-proxy/0.log" Dec 01 16:56:21 crc kubenswrapper[4739]: I1201 16:56:21.486118 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/reloader/0.log" Dec 01 16:56:21 crc kubenswrapper[4739]: I1201 16:56:21.509942 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-zs4ss_d6492935-8d9c-4826-aa9d-26e4e27af9a0/frr-k8s-webhook-server/0.log" Dec 01 16:56:21 crc kubenswrapper[4739]: I1201 16:56:21.860276 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-697d66d89d-jhzwv_c577e6ac-c778-4859-8b39-30aeee31c0ec/manager/0.log" Dec 01 16:56:22 crc kubenswrapper[4739]: I1201 16:56:22.012926 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7cc9fc4f9f-67mlj_1d87d491-a2ce-4749-b3fd-8f429eeeda76/webhook-server/0.log" Dec 01 16:56:22 crc kubenswrapper[4739]: I1201 16:56:22.169958 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cwjct_f3469c68-6d37-41c5-bbca-32d8703b60e4/kube-rbac-proxy/0.log" Dec 01 16:56:22 crc kubenswrapper[4739]: I1201 16:56:22.403853 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hzk9q_6b3db7f6-71d0-4951-89e8-da65693606e7/frr/0.log" Dec 01 16:56:22 crc kubenswrapper[4739]: I1201 16:56:22.562557 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cwjct_f3469c68-6d37-41c5-bbca-32d8703b60e4/speaker/0.log" Dec 01 16:56:36 crc kubenswrapper[4739]: I1201 16:56:36.534070 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp_9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9/util/0.log" Dec 01 16:56:36 crc kubenswrapper[4739]: I1201 16:56:36.696783 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp_9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9/util/0.log" Dec 01 16:56:36 crc kubenswrapper[4739]: I1201 16:56:36.752908 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp_9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9/pull/0.log" Dec 01 16:56:36 crc kubenswrapper[4739]: I1201 16:56:36.766862 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp_9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9/pull/0.log" Dec 01 16:56:36 crc kubenswrapper[4739]: I1201 16:56:36.911073 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp_9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9/util/0.log" Dec 01 16:56:36 crc kubenswrapper[4739]: I1201 16:56:36.922872 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp_9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9/extract/0.log" Dec 01 16:56:36 crc kubenswrapper[4739]: I1201 16:56:36.948555 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9z8dp_9d5ed8aa-55b1-420c-95ca-9f2d2ac571b9/pull/0.log" Dec 01 16:56:37 crc kubenswrapper[4739]: I1201 16:56:37.133203 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s_7beb8fab-f0fc-4d48-8688-ae02488b2e51/util/0.log" Dec 01 16:56:37 crc kubenswrapper[4739]: I1201 16:56:37.285250 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s_7beb8fab-f0fc-4d48-8688-ae02488b2e51/util/0.log" Dec 01 16:56:37 crc kubenswrapper[4739]: I1201 16:56:37.316791 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s_7beb8fab-f0fc-4d48-8688-ae02488b2e51/pull/0.log" Dec 01 16:56:37 crc kubenswrapper[4739]: I1201 16:56:37.339241 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s_7beb8fab-f0fc-4d48-8688-ae02488b2e51/pull/0.log" Dec 01 16:56:37 crc kubenswrapper[4739]: I1201 16:56:37.696225 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s_7beb8fab-f0fc-4d48-8688-ae02488b2e51/util/0.log" Dec 01 16:56:37 crc kubenswrapper[4739]: I1201 16:56:37.770922 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s_7beb8fab-f0fc-4d48-8688-ae02488b2e51/pull/0.log" Dec 01 16:56:37 crc kubenswrapper[4739]: I1201 16:56:37.876972 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8385g2s_7beb8fab-f0fc-4d48-8688-ae02488b2e51/extract/0.log" Dec 01 16:56:37 crc kubenswrapper[4739]: I1201 16:56:37.933025 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xq8q7_d5b77aa5-36d6-4cfd-8742-e71853ba16dd/extract-utilities/0.log" Dec 01 16:56:38 crc kubenswrapper[4739]: I1201 16:56:38.079318 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xq8q7_d5b77aa5-36d6-4cfd-8742-e71853ba16dd/extract-utilities/0.log" Dec 01 16:56:38 crc kubenswrapper[4739]: I1201 16:56:38.116280 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xq8q7_d5b77aa5-36d6-4cfd-8742-e71853ba16dd/extract-content/0.log" Dec 01 16:56:38 crc kubenswrapper[4739]: I1201 16:56:38.134531 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xq8q7_d5b77aa5-36d6-4cfd-8742-e71853ba16dd/extract-content/0.log" Dec 01 16:56:38 crc kubenswrapper[4739]: I1201 16:56:38.295567 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xq8q7_d5b77aa5-36d6-4cfd-8742-e71853ba16dd/extract-utilities/0.log" Dec 01 16:56:38 crc kubenswrapper[4739]: I1201 16:56:38.330777 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xq8q7_d5b77aa5-36d6-4cfd-8742-e71853ba16dd/extract-content/0.log" Dec 01 16:56:38 crc kubenswrapper[4739]: I1201 16:56:38.503047 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hf8ns_cdcee334-c748-423c-ae7b-a4de1795a21b/extract-utilities/0.log" Dec 01 16:56:38 crc kubenswrapper[4739]: I1201 16:56:38.600002 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xq8q7_d5b77aa5-36d6-4cfd-8742-e71853ba16dd/registry-server/0.log" Dec 01 16:56:38 crc kubenswrapper[4739]: I1201 16:56:38.756527 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hf8ns_cdcee334-c748-423c-ae7b-a4de1795a21b/extract-utilities/0.log" Dec 01 16:56:38 crc kubenswrapper[4739]: I1201 16:56:38.757803 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hf8ns_cdcee334-c748-423c-ae7b-a4de1795a21b/extract-content/0.log" Dec 01 16:56:38 crc kubenswrapper[4739]: I1201 16:56:38.782388 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hf8ns_cdcee334-c748-423c-ae7b-a4de1795a21b/extract-content/0.log" Dec 01 16:56:38 crc kubenswrapper[4739]: I1201 16:56:38.951015 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hf8ns_cdcee334-c748-423c-ae7b-a4de1795a21b/extract-utilities/0.log" Dec 01 16:56:39 crc kubenswrapper[4739]: I1201 16:56:39.011883 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hf8ns_cdcee334-c748-423c-ae7b-a4de1795a21b/extract-content/0.log" Dec 01 16:56:39 crc kubenswrapper[4739]: I1201 16:56:39.243050 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hf8ns_cdcee334-c748-423c-ae7b-a4de1795a21b/registry-server/0.log" Dec 01 16:56:39 crc kubenswrapper[4739]: I1201 16:56:39.286862 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-8mjpt_a480b7ca-0552-40d0-ae91-cccfcba31229/marketplace-operator/0.log" Dec 01 16:56:39 crc kubenswrapper[4739]: I1201 16:56:39.321465 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-28lxr_a40475f1-4984-4d14-95ff-9ee7f6493756/extract-utilities/0.log" Dec 01 16:56:39 crc kubenswrapper[4739]: I1201 16:56:39.476348 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-28lxr_a40475f1-4984-4d14-95ff-9ee7f6493756/extract-utilities/0.log" Dec 01 16:56:39 crc kubenswrapper[4739]: I1201 16:56:39.524281 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-28lxr_a40475f1-4984-4d14-95ff-9ee7f6493756/extract-content/0.log" Dec 01 16:56:39 crc kubenswrapper[4739]: I1201 16:56:39.537692 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-28lxr_a40475f1-4984-4d14-95ff-9ee7f6493756/extract-content/0.log" Dec 01 16:56:39 crc kubenswrapper[4739]: I1201 16:56:39.622166 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:56:39 crc kubenswrapper[4739]: I1201 16:56:39.622234 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:56:39 crc kubenswrapper[4739]: I1201 16:56:39.781576 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-28lxr_a40475f1-4984-4d14-95ff-9ee7f6493756/extract-utilities/0.log" Dec 01 16:56:39 crc kubenswrapper[4739]: I1201 16:56:39.813131 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-28lxr_a40475f1-4984-4d14-95ff-9ee7f6493756/extract-content/0.log" Dec 01 16:56:39 crc kubenswrapper[4739]: I1201 16:56:39.894012 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-28lxr_a40475f1-4984-4d14-95ff-9ee7f6493756/registry-server/0.log" Dec 01 16:56:40 crc kubenswrapper[4739]: I1201 16:56:40.018056 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6dj6_658c7094-c779-4d53-bd61-fc0b76e02934/extract-utilities/0.log" Dec 01 16:56:40 crc kubenswrapper[4739]: I1201 16:56:40.207955 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6dj6_658c7094-c779-4d53-bd61-fc0b76e02934/extract-utilities/0.log" Dec 01 16:56:40 crc kubenswrapper[4739]: I1201 16:56:40.263629 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6dj6_658c7094-c779-4d53-bd61-fc0b76e02934/extract-content/0.log" Dec 01 16:56:40 crc kubenswrapper[4739]: I1201 16:56:40.297441 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6dj6_658c7094-c779-4d53-bd61-fc0b76e02934/extract-content/0.log" Dec 01 16:56:40 crc kubenswrapper[4739]: I1201 16:56:40.422396 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6dj6_658c7094-c779-4d53-bd61-fc0b76e02934/extract-utilities/0.log" Dec 01 16:56:40 crc kubenswrapper[4739]: I1201 16:56:40.430946 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6dj6_658c7094-c779-4d53-bd61-fc0b76e02934/extract-content/0.log" Dec 01 16:56:41 crc kubenswrapper[4739]: I1201 16:56:41.465846 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6dj6_658c7094-c779-4d53-bd61-fc0b76e02934/registry-server/0.log" Dec 01 16:56:49 crc kubenswrapper[4739]: I1201 16:56:49.000099 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_d7438aa0-1e45-4112-9af4-a584825d29df/tempest-tests-tempest-tests-runner/0.log" Dec 01 16:57:09 crc kubenswrapper[4739]: I1201 16:57:09.621271 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:57:09 crc kubenswrapper[4739]: I1201 16:57:09.621891 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:57:13 crc kubenswrapper[4739]: E1201 16:57:13.308594 4739 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.106:51586->38.102.83.106:40319: read tcp 38.102.83.106:51586->38.102.83.106:40319: read: connection reset by peer Dec 01 16:57:39 crc kubenswrapper[4739]: I1201 16:57:39.621861 4739 patch_prober.go:28] interesting pod/machine-config-daemon-rnmn7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:57:39 crc kubenswrapper[4739]: I1201 16:57:39.622499 4739 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:57:39 crc kubenswrapper[4739]: I1201 16:57:39.622559 4739 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" Dec 01 16:57:39 crc kubenswrapper[4739]: I1201 16:57:39.623445 4739 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6"} pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:57:39 crc kubenswrapper[4739]: I1201 16:57:39.623548 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerName="machine-config-daemon" containerID="cri-o://d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" gracePeriod=600 Dec 01 16:57:39 crc kubenswrapper[4739]: E1201 16:57:39.754523 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:57:39 crc kubenswrapper[4739]: I1201 16:57:39.777795 4739 generic.go:334] "Generic (PLEG): container finished" podID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" exitCode=0 Dec 01 16:57:39 crc kubenswrapper[4739]: I1201 16:57:39.777828 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerDied","Data":"d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6"} Dec 01 16:57:39 crc kubenswrapper[4739]: I1201 16:57:39.777913 4739 scope.go:117] "RemoveContainer" containerID="af6de8887e1f92dd65adf332c6767d5ff238c9babbbeb1db75ee37509b77d751" Dec 01 16:57:39 crc kubenswrapper[4739]: I1201 16:57:39.778602 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 16:57:39 crc kubenswrapper[4739]: E1201 16:57:39.778897 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:57:54 crc kubenswrapper[4739]: I1201 16:57:54.477789 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 16:57:54 crc kubenswrapper[4739]: E1201 16:57:54.478778 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:58:09 crc kubenswrapper[4739]: I1201 16:58:09.477288 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 16:58:09 crc kubenswrapper[4739]: E1201 16:58:09.477973 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:58:23 crc kubenswrapper[4739]: I1201 16:58:23.478852 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 16:58:23 crc kubenswrapper[4739]: E1201 16:58:23.479773 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:58:34 crc kubenswrapper[4739]: I1201 16:58:34.478571 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 16:58:34 crc kubenswrapper[4739]: E1201 16:58:34.479394 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:58:42 crc kubenswrapper[4739]: I1201 16:58:42.354499 4739 generic.go:334] "Generic (PLEG): container finished" podID="efcb363f-0055-45d3-960a-bb5eef4a7d37" containerID="92cd6803f38c3c5f14919d50392dadf5dad8102307d1caf7aff76d5976f7fa50" exitCode=0 Dec 01 16:58:42 crc kubenswrapper[4739]: I1201 16:58:42.354601 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bsh7j/must-gather-ssxt2" event={"ID":"efcb363f-0055-45d3-960a-bb5eef4a7d37","Type":"ContainerDied","Data":"92cd6803f38c3c5f14919d50392dadf5dad8102307d1caf7aff76d5976f7fa50"} Dec 01 16:58:42 crc kubenswrapper[4739]: I1201 16:58:42.356660 4739 scope.go:117] "RemoveContainer" containerID="92cd6803f38c3c5f14919d50392dadf5dad8102307d1caf7aff76d5976f7fa50" Dec 01 16:58:42 crc kubenswrapper[4739]: I1201 16:58:42.767341 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bsh7j_must-gather-ssxt2_efcb363f-0055-45d3-960a-bb5eef4a7d37/gather/0.log" Dec 01 16:58:47 crc kubenswrapper[4739]: I1201 16:58:47.478115 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 16:58:47 crc kubenswrapper[4739]: E1201 16:58:47.479117 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:58:51 crc kubenswrapper[4739]: I1201 16:58:51.029331 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bsh7j/must-gather-ssxt2"] Dec 01 16:58:51 crc kubenswrapper[4739]: I1201 16:58:51.030056 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-bsh7j/must-gather-ssxt2" podUID="efcb363f-0055-45d3-960a-bb5eef4a7d37" containerName="copy" containerID="cri-o://d8520cd9036bee74c67a88a53baf5c781fea56f100b2745a1800d9a0f2ad770b" gracePeriod=2 Dec 01 16:58:51 crc kubenswrapper[4739]: I1201 16:58:51.052859 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bsh7j/must-gather-ssxt2"] Dec 01 16:58:51 crc kubenswrapper[4739]: I1201 16:58:51.442226 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bsh7j_must-gather-ssxt2_efcb363f-0055-45d3-960a-bb5eef4a7d37/copy/0.log" Dec 01 16:58:51 crc kubenswrapper[4739]: I1201 16:58:51.443072 4739 generic.go:334] "Generic (PLEG): container finished" podID="efcb363f-0055-45d3-960a-bb5eef4a7d37" containerID="d8520cd9036bee74c67a88a53baf5c781fea56f100b2745a1800d9a0f2ad770b" exitCode=143 Dec 01 16:58:51 crc kubenswrapper[4739]: I1201 16:58:51.443173 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb552bc57a850fedaebb5b8788b905015c558d914dafdd77b4ee92428e85b9e2" Dec 01 16:58:51 crc kubenswrapper[4739]: I1201 16:58:51.455660 4739 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bsh7j_must-gather-ssxt2_efcb363f-0055-45d3-960a-bb5eef4a7d37/copy/0.log" Dec 01 16:58:51 crc kubenswrapper[4739]: I1201 16:58:51.456239 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/must-gather-ssxt2" Dec 01 16:58:51 crc kubenswrapper[4739]: I1201 16:58:51.538913 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m49hq\" (UniqueName: \"kubernetes.io/projected/efcb363f-0055-45d3-960a-bb5eef4a7d37-kube-api-access-m49hq\") pod \"efcb363f-0055-45d3-960a-bb5eef4a7d37\" (UID: \"efcb363f-0055-45d3-960a-bb5eef4a7d37\") " Dec 01 16:58:51 crc kubenswrapper[4739]: I1201 16:58:51.539245 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/efcb363f-0055-45d3-960a-bb5eef4a7d37-must-gather-output\") pod \"efcb363f-0055-45d3-960a-bb5eef4a7d37\" (UID: \"efcb363f-0055-45d3-960a-bb5eef4a7d37\") " Dec 01 16:58:51 crc kubenswrapper[4739]: I1201 16:58:51.545647 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efcb363f-0055-45d3-960a-bb5eef4a7d37-kube-api-access-m49hq" (OuterVolumeSpecName: "kube-api-access-m49hq") pod "efcb363f-0055-45d3-960a-bb5eef4a7d37" (UID: "efcb363f-0055-45d3-960a-bb5eef4a7d37"). InnerVolumeSpecName "kube-api-access-m49hq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:58:51 crc kubenswrapper[4739]: I1201 16:58:51.641168 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m49hq\" (UniqueName: \"kubernetes.io/projected/efcb363f-0055-45d3-960a-bb5eef4a7d37-kube-api-access-m49hq\") on node \"crc\" DevicePath \"\"" Dec 01 16:58:51 crc kubenswrapper[4739]: I1201 16:58:51.734980 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efcb363f-0055-45d3-960a-bb5eef4a7d37-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "efcb363f-0055-45d3-960a-bb5eef4a7d37" (UID: "efcb363f-0055-45d3-960a-bb5eef4a7d37"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:58:51 crc kubenswrapper[4739]: I1201 16:58:51.743020 4739 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/efcb363f-0055-45d3-960a-bb5eef4a7d37-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 01 16:58:52 crc kubenswrapper[4739]: I1201 16:58:52.451520 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bsh7j/must-gather-ssxt2" Dec 01 16:58:52 crc kubenswrapper[4739]: I1201 16:58:52.490785 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efcb363f-0055-45d3-960a-bb5eef4a7d37" path="/var/lib/kubelet/pods/efcb363f-0055-45d3-960a-bb5eef4a7d37/volumes" Dec 01 16:58:59 crc kubenswrapper[4739]: I1201 16:58:59.478129 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 16:58:59 crc kubenswrapper[4739]: E1201 16:58:59.478906 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:59:11 crc kubenswrapper[4739]: I1201 16:59:11.477196 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 16:59:11 crc kubenswrapper[4739]: E1201 16:59:11.478081 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:59:22 crc kubenswrapper[4739]: I1201 16:59:22.479443 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 16:59:22 crc kubenswrapper[4739]: E1201 16:59:22.481760 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:59:25 crc kubenswrapper[4739]: I1201 16:59:25.622893 4739 scope.go:117] "RemoveContainer" containerID="d8520cd9036bee74c67a88a53baf5c781fea56f100b2745a1800d9a0f2ad770b" Dec 01 16:59:25 crc kubenswrapper[4739]: I1201 16:59:25.669179 4739 scope.go:117] "RemoveContainer" containerID="92cd6803f38c3c5f14919d50392dadf5dad8102307d1caf7aff76d5976f7fa50" Dec 01 16:59:33 crc kubenswrapper[4739]: I1201 16:59:33.477668 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 16:59:33 crc kubenswrapper[4739]: E1201 16:59:33.478531 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:59:47 crc kubenswrapper[4739]: I1201 16:59:47.012583 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 16:59:47 crc kubenswrapper[4739]: E1201 16:59:47.013300 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 16:59:59 crc kubenswrapper[4739]: I1201 16:59:59.476870 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 16:59:59 crc kubenswrapper[4739]: E1201 16:59:59.477880 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.157463 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56"] Dec 01 17:00:00 crc kubenswrapper[4739]: E1201 17:00:00.157877 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efcb363f-0055-45d3-960a-bb5eef4a7d37" containerName="gather" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.157890 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="efcb363f-0055-45d3-960a-bb5eef4a7d37" containerName="gather" Dec 01 17:00:00 crc kubenswrapper[4739]: E1201 17:00:00.157902 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efcb363f-0055-45d3-960a-bb5eef4a7d37" containerName="copy" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.157907 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="efcb363f-0055-45d3-960a-bb5eef4a7d37" containerName="copy" Dec 01 17:00:00 crc kubenswrapper[4739]: E1201 17:00:00.157917 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="632e0bd0-9997-470e-81b2-88475ce103f0" containerName="registry-server" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.157923 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="632e0bd0-9997-470e-81b2-88475ce103f0" containerName="registry-server" Dec 01 17:00:00 crc kubenswrapper[4739]: E1201 17:00:00.157934 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="632e0bd0-9997-470e-81b2-88475ce103f0" containerName="extract-utilities" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.157940 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="632e0bd0-9997-470e-81b2-88475ce103f0" containerName="extract-utilities" Dec 01 17:00:00 crc kubenswrapper[4739]: E1201 17:00:00.157951 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="632e0bd0-9997-470e-81b2-88475ce103f0" containerName="extract-content" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.157956 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="632e0bd0-9997-470e-81b2-88475ce103f0" containerName="extract-content" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.158143 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="efcb363f-0055-45d3-960a-bb5eef4a7d37" containerName="gather" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.158155 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="632e0bd0-9997-470e-81b2-88475ce103f0" containerName="registry-server" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.158172 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="efcb363f-0055-45d3-960a-bb5eef4a7d37" containerName="copy" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.158999 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.168293 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56"] Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.199984 4739 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.200523 4739 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.302517 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e376af36-e63c-4b17-8756-72224cc3e975-config-volume\") pod \"collect-profiles-29410140-4rw56\" (UID: \"e376af36-e63c-4b17-8756-72224cc3e975\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.302574 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e376af36-e63c-4b17-8756-72224cc3e975-secret-volume\") pod \"collect-profiles-29410140-4rw56\" (UID: \"e376af36-e63c-4b17-8756-72224cc3e975\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.303158 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twlfx\" (UniqueName: \"kubernetes.io/projected/e376af36-e63c-4b17-8756-72224cc3e975-kube-api-access-twlfx\") pod \"collect-profiles-29410140-4rw56\" (UID: \"e376af36-e63c-4b17-8756-72224cc3e975\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.404931 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e376af36-e63c-4b17-8756-72224cc3e975-config-volume\") pod \"collect-profiles-29410140-4rw56\" (UID: \"e376af36-e63c-4b17-8756-72224cc3e975\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.404993 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e376af36-e63c-4b17-8756-72224cc3e975-secret-volume\") pod \"collect-profiles-29410140-4rw56\" (UID: \"e376af36-e63c-4b17-8756-72224cc3e975\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.405064 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twlfx\" (UniqueName: \"kubernetes.io/projected/e376af36-e63c-4b17-8756-72224cc3e975-kube-api-access-twlfx\") pod \"collect-profiles-29410140-4rw56\" (UID: \"e376af36-e63c-4b17-8756-72224cc3e975\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.405917 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e376af36-e63c-4b17-8756-72224cc3e975-config-volume\") pod \"collect-profiles-29410140-4rw56\" (UID: \"e376af36-e63c-4b17-8756-72224cc3e975\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.423752 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e376af36-e63c-4b17-8756-72224cc3e975-secret-volume\") pod \"collect-profiles-29410140-4rw56\" (UID: \"e376af36-e63c-4b17-8756-72224cc3e975\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.440121 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twlfx\" (UniqueName: \"kubernetes.io/projected/e376af36-e63c-4b17-8756-72224cc3e975-kube-api-access-twlfx\") pod \"collect-profiles-29410140-4rw56\" (UID: \"e376af36-e63c-4b17-8756-72224cc3e975\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.522594 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" Dec 01 17:00:00 crc kubenswrapper[4739]: I1201 17:00:00.994134 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56"] Dec 01 17:00:01 crc kubenswrapper[4739]: W1201 17:00:01.012916 4739 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode376af36_e63c_4b17_8756_72224cc3e975.slice/crio-b56e2531ed61d1bcb2bb6f415aa33c52b454cfb2e1ac8327847887870789c3f6 WatchSource:0}: Error finding container b56e2531ed61d1bcb2bb6f415aa33c52b454cfb2e1ac8327847887870789c3f6: Status 404 returned error can't find the container with id b56e2531ed61d1bcb2bb6f415aa33c52b454cfb2e1ac8327847887870789c3f6 Dec 01 17:00:01 crc kubenswrapper[4739]: I1201 17:00:01.214999 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" event={"ID":"e376af36-e63c-4b17-8756-72224cc3e975","Type":"ContainerStarted","Data":"8b20f81bb384010903013b8f146a07a06eca32e8bf52928180377ed76a3eda9a"} Dec 01 17:00:01 crc kubenswrapper[4739]: I1201 17:00:01.215048 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" event={"ID":"e376af36-e63c-4b17-8756-72224cc3e975","Type":"ContainerStarted","Data":"b56e2531ed61d1bcb2bb6f415aa33c52b454cfb2e1ac8327847887870789c3f6"} Dec 01 17:00:02 crc kubenswrapper[4739]: I1201 17:00:02.231035 4739 generic.go:334] "Generic (PLEG): container finished" podID="e376af36-e63c-4b17-8756-72224cc3e975" containerID="8b20f81bb384010903013b8f146a07a06eca32e8bf52928180377ed76a3eda9a" exitCode=0 Dec 01 17:00:02 crc kubenswrapper[4739]: I1201 17:00:02.231165 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" event={"ID":"e376af36-e63c-4b17-8756-72224cc3e975","Type":"ContainerDied","Data":"8b20f81bb384010903013b8f146a07a06eca32e8bf52928180377ed76a3eda9a"} Dec 01 17:00:03 crc kubenswrapper[4739]: I1201 17:00:03.552851 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" Dec 01 17:00:03 crc kubenswrapper[4739]: I1201 17:00:03.664168 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e376af36-e63c-4b17-8756-72224cc3e975-config-volume\") pod \"e376af36-e63c-4b17-8756-72224cc3e975\" (UID: \"e376af36-e63c-4b17-8756-72224cc3e975\") " Dec 01 17:00:03 crc kubenswrapper[4739]: I1201 17:00:03.664654 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twlfx\" (UniqueName: \"kubernetes.io/projected/e376af36-e63c-4b17-8756-72224cc3e975-kube-api-access-twlfx\") pod \"e376af36-e63c-4b17-8756-72224cc3e975\" (UID: \"e376af36-e63c-4b17-8756-72224cc3e975\") " Dec 01 17:00:03 crc kubenswrapper[4739]: I1201 17:00:03.664812 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e376af36-e63c-4b17-8756-72224cc3e975-secret-volume\") pod \"e376af36-e63c-4b17-8756-72224cc3e975\" (UID: \"e376af36-e63c-4b17-8756-72224cc3e975\") " Dec 01 17:00:03 crc kubenswrapper[4739]: I1201 17:00:03.665207 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e376af36-e63c-4b17-8756-72224cc3e975-config-volume" (OuterVolumeSpecName: "config-volume") pod "e376af36-e63c-4b17-8756-72224cc3e975" (UID: "e376af36-e63c-4b17-8756-72224cc3e975"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:00:03 crc kubenswrapper[4739]: I1201 17:00:03.665802 4739 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e376af36-e63c-4b17-8756-72224cc3e975-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 17:00:03 crc kubenswrapper[4739]: I1201 17:00:03.670299 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e376af36-e63c-4b17-8756-72224cc3e975-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e376af36-e63c-4b17-8756-72224cc3e975" (UID: "e376af36-e63c-4b17-8756-72224cc3e975"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:00:03 crc kubenswrapper[4739]: I1201 17:00:03.670355 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e376af36-e63c-4b17-8756-72224cc3e975-kube-api-access-twlfx" (OuterVolumeSpecName: "kube-api-access-twlfx") pod "e376af36-e63c-4b17-8756-72224cc3e975" (UID: "e376af36-e63c-4b17-8756-72224cc3e975"). InnerVolumeSpecName "kube-api-access-twlfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:00:03 crc kubenswrapper[4739]: I1201 17:00:03.766902 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twlfx\" (UniqueName: \"kubernetes.io/projected/e376af36-e63c-4b17-8756-72224cc3e975-kube-api-access-twlfx\") on node \"crc\" DevicePath \"\"" Dec 01 17:00:03 crc kubenswrapper[4739]: I1201 17:00:03.767202 4739 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e376af36-e63c-4b17-8756-72224cc3e975-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 17:00:04 crc kubenswrapper[4739]: I1201 17:00:04.272371 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" event={"ID":"e376af36-e63c-4b17-8756-72224cc3e975","Type":"ContainerDied","Data":"b56e2531ed61d1bcb2bb6f415aa33c52b454cfb2e1ac8327847887870789c3f6"} Dec 01 17:00:04 crc kubenswrapper[4739]: I1201 17:00:04.272430 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b56e2531ed61d1bcb2bb6f415aa33c52b454cfb2e1ac8327847887870789c3f6" Dec 01 17:00:04 crc kubenswrapper[4739]: I1201 17:00:04.272487 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-4rw56" Dec 01 17:00:04 crc kubenswrapper[4739]: I1201 17:00:04.622704 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn"] Dec 01 17:00:04 crc kubenswrapper[4739]: I1201 17:00:04.630395 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410095-z4gfn"] Dec 01 17:00:06 crc kubenswrapper[4739]: I1201 17:00:06.492060 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63ae283a-eeca-4a87-baad-8ab09d916f1e" path="/var/lib/kubelet/pods/63ae283a-eeca-4a87-baad-8ab09d916f1e/volumes" Dec 01 17:00:09 crc kubenswrapper[4739]: I1201 17:00:09.639370 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-whp5h"] Dec 01 17:00:09 crc kubenswrapper[4739]: E1201 17:00:09.643895 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e376af36-e63c-4b17-8756-72224cc3e975" containerName="collect-profiles" Dec 01 17:00:09 crc kubenswrapper[4739]: I1201 17:00:09.643920 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="e376af36-e63c-4b17-8756-72224cc3e975" containerName="collect-profiles" Dec 01 17:00:09 crc kubenswrapper[4739]: I1201 17:00:09.644182 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="e376af36-e63c-4b17-8756-72224cc3e975" containerName="collect-profiles" Dec 01 17:00:09 crc kubenswrapper[4739]: I1201 17:00:09.646336 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:09 crc kubenswrapper[4739]: I1201 17:00:09.652721 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-whp5h"] Dec 01 17:00:09 crc kubenswrapper[4739]: I1201 17:00:09.677174 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtprf\" (UniqueName: \"kubernetes.io/projected/7348b244-3ab0-465a-8cfd-82db16f4b8c3-kube-api-access-wtprf\") pod \"community-operators-whp5h\" (UID: \"7348b244-3ab0-465a-8cfd-82db16f4b8c3\") " pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:09 crc kubenswrapper[4739]: I1201 17:00:09.677227 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7348b244-3ab0-465a-8cfd-82db16f4b8c3-catalog-content\") pod \"community-operators-whp5h\" (UID: \"7348b244-3ab0-465a-8cfd-82db16f4b8c3\") " pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:09 crc kubenswrapper[4739]: I1201 17:00:09.677437 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7348b244-3ab0-465a-8cfd-82db16f4b8c3-utilities\") pod \"community-operators-whp5h\" (UID: \"7348b244-3ab0-465a-8cfd-82db16f4b8c3\") " pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:09 crc kubenswrapper[4739]: I1201 17:00:09.779053 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7348b244-3ab0-465a-8cfd-82db16f4b8c3-utilities\") pod \"community-operators-whp5h\" (UID: \"7348b244-3ab0-465a-8cfd-82db16f4b8c3\") " pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:09 crc kubenswrapper[4739]: I1201 17:00:09.779228 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtprf\" (UniqueName: \"kubernetes.io/projected/7348b244-3ab0-465a-8cfd-82db16f4b8c3-kube-api-access-wtprf\") pod \"community-operators-whp5h\" (UID: \"7348b244-3ab0-465a-8cfd-82db16f4b8c3\") " pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:09 crc kubenswrapper[4739]: I1201 17:00:09.779278 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7348b244-3ab0-465a-8cfd-82db16f4b8c3-catalog-content\") pod \"community-operators-whp5h\" (UID: \"7348b244-3ab0-465a-8cfd-82db16f4b8c3\") " pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:09 crc kubenswrapper[4739]: I1201 17:00:09.779976 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7348b244-3ab0-465a-8cfd-82db16f4b8c3-utilities\") pod \"community-operators-whp5h\" (UID: \"7348b244-3ab0-465a-8cfd-82db16f4b8c3\") " pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:09 crc kubenswrapper[4739]: I1201 17:00:09.780038 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7348b244-3ab0-465a-8cfd-82db16f4b8c3-catalog-content\") pod \"community-operators-whp5h\" (UID: \"7348b244-3ab0-465a-8cfd-82db16f4b8c3\") " pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:09 crc kubenswrapper[4739]: I1201 17:00:09.808602 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtprf\" (UniqueName: \"kubernetes.io/projected/7348b244-3ab0-465a-8cfd-82db16f4b8c3-kube-api-access-wtprf\") pod \"community-operators-whp5h\" (UID: \"7348b244-3ab0-465a-8cfd-82db16f4b8c3\") " pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:09 crc kubenswrapper[4739]: I1201 17:00:09.994846 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:10 crc kubenswrapper[4739]: I1201 17:00:10.372024 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-whp5h"] Dec 01 17:00:10 crc kubenswrapper[4739]: I1201 17:00:10.478638 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 17:00:10 crc kubenswrapper[4739]: E1201 17:00:10.479124 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 17:00:11 crc kubenswrapper[4739]: I1201 17:00:11.332941 4739 generic.go:334] "Generic (PLEG): container finished" podID="7348b244-3ab0-465a-8cfd-82db16f4b8c3" containerID="01283a7d5ad5d369cfe7ae4bcc8dd380628fa37d0e2a5530be531294646ba78a" exitCode=0 Dec 01 17:00:11 crc kubenswrapper[4739]: I1201 17:00:11.332985 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whp5h" event={"ID":"7348b244-3ab0-465a-8cfd-82db16f4b8c3","Type":"ContainerDied","Data":"01283a7d5ad5d369cfe7ae4bcc8dd380628fa37d0e2a5530be531294646ba78a"} Dec 01 17:00:11 crc kubenswrapper[4739]: I1201 17:00:11.333008 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whp5h" event={"ID":"7348b244-3ab0-465a-8cfd-82db16f4b8c3","Type":"ContainerStarted","Data":"5cb3d78c899a825f8f1bf248350320926e4b69d85a2fb75b4eeeb8f3ecf1fe1d"} Dec 01 17:00:12 crc kubenswrapper[4739]: I1201 17:00:12.343579 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whp5h" event={"ID":"7348b244-3ab0-465a-8cfd-82db16f4b8c3","Type":"ContainerStarted","Data":"5a2dc06926139c4777b4f2cc30c3d1c96da53f349f4ebd08461f7db72467ac7f"} Dec 01 17:00:13 crc kubenswrapper[4739]: I1201 17:00:13.354178 4739 generic.go:334] "Generic (PLEG): container finished" podID="7348b244-3ab0-465a-8cfd-82db16f4b8c3" containerID="5a2dc06926139c4777b4f2cc30c3d1c96da53f349f4ebd08461f7db72467ac7f" exitCode=0 Dec 01 17:00:13 crc kubenswrapper[4739]: I1201 17:00:13.354237 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whp5h" event={"ID":"7348b244-3ab0-465a-8cfd-82db16f4b8c3","Type":"ContainerDied","Data":"5a2dc06926139c4777b4f2cc30c3d1c96da53f349f4ebd08461f7db72467ac7f"} Dec 01 17:00:14 crc kubenswrapper[4739]: I1201 17:00:14.366045 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whp5h" event={"ID":"7348b244-3ab0-465a-8cfd-82db16f4b8c3","Type":"ContainerStarted","Data":"67d10774daec8e833549652fd9ff4ab19d929e8bdab534f723f3d439660c009d"} Dec 01 17:00:14 crc kubenswrapper[4739]: I1201 17:00:14.385938 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-whp5h" podStartSLOduration=2.696635298 podStartE2EDuration="5.385920747s" podCreationTimestamp="2025-12-01 17:00:09 +0000 UTC" firstStartedPulling="2025-12-01 17:00:11.334607425 +0000 UTC m=+5113.160353519" lastFinishedPulling="2025-12-01 17:00:14.023892874 +0000 UTC m=+5115.849638968" observedRunningTime="2025-12-01 17:00:14.382171221 +0000 UTC m=+5116.207917335" watchObservedRunningTime="2025-12-01 17:00:14.385920747 +0000 UTC m=+5116.211666841" Dec 01 17:00:19 crc kubenswrapper[4739]: I1201 17:00:19.995019 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:19 crc kubenswrapper[4739]: I1201 17:00:19.995871 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:20 crc kubenswrapper[4739]: I1201 17:00:20.302466 4739 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:20 crc kubenswrapper[4739]: I1201 17:00:20.470155 4739 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:22 crc kubenswrapper[4739]: I1201 17:00:22.332868 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-whp5h"] Dec 01 17:00:22 crc kubenswrapper[4739]: I1201 17:00:22.442381 4739 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-whp5h" podUID="7348b244-3ab0-465a-8cfd-82db16f4b8c3" containerName="registry-server" containerID="cri-o://67d10774daec8e833549652fd9ff4ab19d929e8bdab534f723f3d439660c009d" gracePeriod=2 Dec 01 17:00:22 crc kubenswrapper[4739]: I1201 17:00:22.476962 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 17:00:22 crc kubenswrapper[4739]: E1201 17:00:22.477316 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 17:00:22 crc kubenswrapper[4739]: I1201 17:00:22.934453 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.054658 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtprf\" (UniqueName: \"kubernetes.io/projected/7348b244-3ab0-465a-8cfd-82db16f4b8c3-kube-api-access-wtprf\") pod \"7348b244-3ab0-465a-8cfd-82db16f4b8c3\" (UID: \"7348b244-3ab0-465a-8cfd-82db16f4b8c3\") " Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.054799 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7348b244-3ab0-465a-8cfd-82db16f4b8c3-utilities\") pod \"7348b244-3ab0-465a-8cfd-82db16f4b8c3\" (UID: \"7348b244-3ab0-465a-8cfd-82db16f4b8c3\") " Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.054851 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7348b244-3ab0-465a-8cfd-82db16f4b8c3-catalog-content\") pod \"7348b244-3ab0-465a-8cfd-82db16f4b8c3\" (UID: \"7348b244-3ab0-465a-8cfd-82db16f4b8c3\") " Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.055936 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7348b244-3ab0-465a-8cfd-82db16f4b8c3-utilities" (OuterVolumeSpecName: "utilities") pod "7348b244-3ab0-465a-8cfd-82db16f4b8c3" (UID: "7348b244-3ab0-465a-8cfd-82db16f4b8c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.060589 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7348b244-3ab0-465a-8cfd-82db16f4b8c3-kube-api-access-wtprf" (OuterVolumeSpecName: "kube-api-access-wtprf") pod "7348b244-3ab0-465a-8cfd-82db16f4b8c3" (UID: "7348b244-3ab0-465a-8cfd-82db16f4b8c3"). InnerVolumeSpecName "kube-api-access-wtprf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.157770 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtprf\" (UniqueName: \"kubernetes.io/projected/7348b244-3ab0-465a-8cfd-82db16f4b8c3-kube-api-access-wtprf\") on node \"crc\" DevicePath \"\"" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.157818 4739 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7348b244-3ab0-465a-8cfd-82db16f4b8c3-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.177172 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7348b244-3ab0-465a-8cfd-82db16f4b8c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7348b244-3ab0-465a-8cfd-82db16f4b8c3" (UID: "7348b244-3ab0-465a-8cfd-82db16f4b8c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.260248 4739 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7348b244-3ab0-465a-8cfd-82db16f4b8c3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.451565 4739 generic.go:334] "Generic (PLEG): container finished" podID="7348b244-3ab0-465a-8cfd-82db16f4b8c3" containerID="67d10774daec8e833549652fd9ff4ab19d929e8bdab534f723f3d439660c009d" exitCode=0 Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.451604 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whp5h" event={"ID":"7348b244-3ab0-465a-8cfd-82db16f4b8c3","Type":"ContainerDied","Data":"67d10774daec8e833549652fd9ff4ab19d929e8bdab534f723f3d439660c009d"} Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.451628 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whp5h" event={"ID":"7348b244-3ab0-465a-8cfd-82db16f4b8c3","Type":"ContainerDied","Data":"5cb3d78c899a825f8f1bf248350320926e4b69d85a2fb75b4eeeb8f3ecf1fe1d"} Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.451645 4739 scope.go:117] "RemoveContainer" containerID="67d10774daec8e833549652fd9ff4ab19d929e8bdab534f723f3d439660c009d" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.451769 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whp5h" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.485001 4739 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-whp5h"] Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.486498 4739 scope.go:117] "RemoveContainer" containerID="5a2dc06926139c4777b4f2cc30c3d1c96da53f349f4ebd08461f7db72467ac7f" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.504943 4739 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-whp5h"] Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.520710 4739 scope.go:117] "RemoveContainer" containerID="01283a7d5ad5d369cfe7ae4bcc8dd380628fa37d0e2a5530be531294646ba78a" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.555755 4739 scope.go:117] "RemoveContainer" containerID="67d10774daec8e833549652fd9ff4ab19d929e8bdab534f723f3d439660c009d" Dec 01 17:00:23 crc kubenswrapper[4739]: E1201 17:00:23.556743 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67d10774daec8e833549652fd9ff4ab19d929e8bdab534f723f3d439660c009d\": container with ID starting with 67d10774daec8e833549652fd9ff4ab19d929e8bdab534f723f3d439660c009d not found: ID does not exist" containerID="67d10774daec8e833549652fd9ff4ab19d929e8bdab534f723f3d439660c009d" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.556825 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67d10774daec8e833549652fd9ff4ab19d929e8bdab534f723f3d439660c009d"} err="failed to get container status \"67d10774daec8e833549652fd9ff4ab19d929e8bdab534f723f3d439660c009d\": rpc error: code = NotFound desc = could not find container \"67d10774daec8e833549652fd9ff4ab19d929e8bdab534f723f3d439660c009d\": container with ID starting with 67d10774daec8e833549652fd9ff4ab19d929e8bdab534f723f3d439660c009d not found: ID does not exist" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.556861 4739 scope.go:117] "RemoveContainer" containerID="5a2dc06926139c4777b4f2cc30c3d1c96da53f349f4ebd08461f7db72467ac7f" Dec 01 17:00:23 crc kubenswrapper[4739]: E1201 17:00:23.557679 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a2dc06926139c4777b4f2cc30c3d1c96da53f349f4ebd08461f7db72467ac7f\": container with ID starting with 5a2dc06926139c4777b4f2cc30c3d1c96da53f349f4ebd08461f7db72467ac7f not found: ID does not exist" containerID="5a2dc06926139c4777b4f2cc30c3d1c96da53f349f4ebd08461f7db72467ac7f" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.557742 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a2dc06926139c4777b4f2cc30c3d1c96da53f349f4ebd08461f7db72467ac7f"} err="failed to get container status \"5a2dc06926139c4777b4f2cc30c3d1c96da53f349f4ebd08461f7db72467ac7f\": rpc error: code = NotFound desc = could not find container \"5a2dc06926139c4777b4f2cc30c3d1c96da53f349f4ebd08461f7db72467ac7f\": container with ID starting with 5a2dc06926139c4777b4f2cc30c3d1c96da53f349f4ebd08461f7db72467ac7f not found: ID does not exist" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.557776 4739 scope.go:117] "RemoveContainer" containerID="01283a7d5ad5d369cfe7ae4bcc8dd380628fa37d0e2a5530be531294646ba78a" Dec 01 17:00:23 crc kubenswrapper[4739]: E1201 17:00:23.558575 4739 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01283a7d5ad5d369cfe7ae4bcc8dd380628fa37d0e2a5530be531294646ba78a\": container with ID starting with 01283a7d5ad5d369cfe7ae4bcc8dd380628fa37d0e2a5530be531294646ba78a not found: ID does not exist" containerID="01283a7d5ad5d369cfe7ae4bcc8dd380628fa37d0e2a5530be531294646ba78a" Dec 01 17:00:23 crc kubenswrapper[4739]: I1201 17:00:23.558659 4739 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01283a7d5ad5d369cfe7ae4bcc8dd380628fa37d0e2a5530be531294646ba78a"} err="failed to get container status \"01283a7d5ad5d369cfe7ae4bcc8dd380628fa37d0e2a5530be531294646ba78a\": rpc error: code = NotFound desc = could not find container \"01283a7d5ad5d369cfe7ae4bcc8dd380628fa37d0e2a5530be531294646ba78a\": container with ID starting with 01283a7d5ad5d369cfe7ae4bcc8dd380628fa37d0e2a5530be531294646ba78a not found: ID does not exist" Dec 01 17:00:24 crc kubenswrapper[4739]: I1201 17:00:24.490566 4739 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7348b244-3ab0-465a-8cfd-82db16f4b8c3" path="/var/lib/kubelet/pods/7348b244-3ab0-465a-8cfd-82db16f4b8c3/volumes" Dec 01 17:00:25 crc kubenswrapper[4739]: I1201 17:00:25.778030 4739 scope.go:117] "RemoveContainer" containerID="0ba245c11fbd431a71abab2e4668376b895d1573285b82b9f5dafb2b2e88aef7" Dec 01 17:00:25 crc kubenswrapper[4739]: I1201 17:00:25.803223 4739 scope.go:117] "RemoveContainer" containerID="2fe943a034e9d0086fc3b5e76fb58793cc0449ad65bc3a426e00ca6b6da6a0cc" Dec 01 17:00:34 crc kubenswrapper[4739]: I1201 17:00:34.477351 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 17:00:34 crc kubenswrapper[4739]: E1201 17:00:34.478034 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 17:00:49 crc kubenswrapper[4739]: I1201 17:00:49.478175 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 17:00:49 crc kubenswrapper[4739]: E1201 17:00:49.478998 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.147894 4739 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29410141-jzdc8"] Dec 01 17:01:00 crc kubenswrapper[4739]: E1201 17:01:00.148826 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7348b244-3ab0-465a-8cfd-82db16f4b8c3" containerName="registry-server" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.148839 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7348b244-3ab0-465a-8cfd-82db16f4b8c3" containerName="registry-server" Dec 01 17:01:00 crc kubenswrapper[4739]: E1201 17:01:00.148854 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7348b244-3ab0-465a-8cfd-82db16f4b8c3" containerName="extract-utilities" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.148861 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7348b244-3ab0-465a-8cfd-82db16f4b8c3" containerName="extract-utilities" Dec 01 17:01:00 crc kubenswrapper[4739]: E1201 17:01:00.148875 4739 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7348b244-3ab0-465a-8cfd-82db16f4b8c3" containerName="extract-content" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.148881 4739 state_mem.go:107] "Deleted CPUSet assignment" podUID="7348b244-3ab0-465a-8cfd-82db16f4b8c3" containerName="extract-content" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.149060 4739 memory_manager.go:354] "RemoveStaleState removing state" podUID="7348b244-3ab0-465a-8cfd-82db16f4b8c3" containerName="registry-server" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.149904 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.178016 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29410141-jzdc8"] Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.260247 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-config-data\") pod \"keystone-cron-29410141-jzdc8\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.260338 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk84f\" (UniqueName: \"kubernetes.io/projected/3e1a98b7-aa24-4809-b9a5-ca9be3472781-kube-api-access-vk84f\") pod \"keystone-cron-29410141-jzdc8\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.260522 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-combined-ca-bundle\") pod \"keystone-cron-29410141-jzdc8\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.260591 4739 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-fernet-keys\") pod \"keystone-cron-29410141-jzdc8\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.362337 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-combined-ca-bundle\") pod \"keystone-cron-29410141-jzdc8\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.362455 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-fernet-keys\") pod \"keystone-cron-29410141-jzdc8\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.362529 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-config-data\") pod \"keystone-cron-29410141-jzdc8\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.362586 4739 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk84f\" (UniqueName: \"kubernetes.io/projected/3e1a98b7-aa24-4809-b9a5-ca9be3472781-kube-api-access-vk84f\") pod \"keystone-cron-29410141-jzdc8\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.368571 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-config-data\") pod \"keystone-cron-29410141-jzdc8\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.384826 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk84f\" (UniqueName: \"kubernetes.io/projected/3e1a98b7-aa24-4809-b9a5-ca9be3472781-kube-api-access-vk84f\") pod \"keystone-cron-29410141-jzdc8\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.386032 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-combined-ca-bundle\") pod \"keystone-cron-29410141-jzdc8\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.394934 4739 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-fernet-keys\") pod \"keystone-cron-29410141-jzdc8\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.475073 4739 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:00 crc kubenswrapper[4739]: I1201 17:01:00.978282 4739 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29410141-jzdc8"] Dec 01 17:01:01 crc kubenswrapper[4739]: I1201 17:01:01.945862 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410141-jzdc8" event={"ID":"3e1a98b7-aa24-4809-b9a5-ca9be3472781","Type":"ContainerStarted","Data":"9a22a3acc9d06e8cf9ac9c38944ad5cbef9159505aa176b1540e1d0b645a811c"} Dec 01 17:01:01 crc kubenswrapper[4739]: I1201 17:01:01.946210 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410141-jzdc8" event={"ID":"3e1a98b7-aa24-4809-b9a5-ca9be3472781","Type":"ContainerStarted","Data":"8a43e23df0df1d5156d83f27ce9017231103590eb4ccf907b8924c4d0eea04fc"} Dec 01 17:01:01 crc kubenswrapper[4739]: I1201 17:01:01.972824 4739 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29410141-jzdc8" podStartSLOduration=1.972807227 podStartE2EDuration="1.972807227s" podCreationTimestamp="2025-12-01 17:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:01:01.96322227 +0000 UTC m=+5163.788968364" watchObservedRunningTime="2025-12-01 17:01:01.972807227 +0000 UTC m=+5163.798553321" Dec 01 17:01:04 crc kubenswrapper[4739]: I1201 17:01:04.477222 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 17:01:04 crc kubenswrapper[4739]: E1201 17:01:04.478016 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 17:01:04 crc kubenswrapper[4739]: I1201 17:01:04.972585 4739 generic.go:334] "Generic (PLEG): container finished" podID="3e1a98b7-aa24-4809-b9a5-ca9be3472781" containerID="9a22a3acc9d06e8cf9ac9c38944ad5cbef9159505aa176b1540e1d0b645a811c" exitCode=0 Dec 01 17:01:04 crc kubenswrapper[4739]: I1201 17:01:04.972663 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410141-jzdc8" event={"ID":"3e1a98b7-aa24-4809-b9a5-ca9be3472781","Type":"ContainerDied","Data":"9a22a3acc9d06e8cf9ac9c38944ad5cbef9159505aa176b1540e1d0b645a811c"} Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.542902 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.618625 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vk84f\" (UniqueName: \"kubernetes.io/projected/3e1a98b7-aa24-4809-b9a5-ca9be3472781-kube-api-access-vk84f\") pod \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.619790 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-config-data\") pod \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.619917 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-combined-ca-bundle\") pod \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.622797 4739 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-fernet-keys\") pod \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\" (UID: \"3e1a98b7-aa24-4809-b9a5-ca9be3472781\") " Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.625656 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e1a98b7-aa24-4809-b9a5-ca9be3472781-kube-api-access-vk84f" (OuterVolumeSpecName: "kube-api-access-vk84f") pod "3e1a98b7-aa24-4809-b9a5-ca9be3472781" (UID: "3e1a98b7-aa24-4809-b9a5-ca9be3472781"). InnerVolumeSpecName "kube-api-access-vk84f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.629330 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3e1a98b7-aa24-4809-b9a5-ca9be3472781" (UID: "3e1a98b7-aa24-4809-b9a5-ca9be3472781"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.657578 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e1a98b7-aa24-4809-b9a5-ca9be3472781" (UID: "3e1a98b7-aa24-4809-b9a5-ca9be3472781"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.683199 4739 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-config-data" (OuterVolumeSpecName: "config-data") pod "3e1a98b7-aa24-4809-b9a5-ca9be3472781" (UID: "3e1a98b7-aa24-4809-b9a5-ca9be3472781"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.726527 4739 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.726575 4739 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.726587 4739 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vk84f\" (UniqueName: \"kubernetes.io/projected/3e1a98b7-aa24-4809-b9a5-ca9be3472781-kube-api-access-vk84f\") on node \"crc\" DevicePath \"\"" Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.726598 4739 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1a98b7-aa24-4809-b9a5-ca9be3472781-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.989191 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410141-jzdc8" event={"ID":"3e1a98b7-aa24-4809-b9a5-ca9be3472781","Type":"ContainerDied","Data":"8a43e23df0df1d5156d83f27ce9017231103590eb4ccf907b8924c4d0eea04fc"} Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.989235 4739 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410141-jzdc8" Dec 01 17:01:06 crc kubenswrapper[4739]: I1201 17:01:06.989245 4739 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a43e23df0df1d5156d83f27ce9017231103590eb4ccf907b8924c4d0eea04fc" Dec 01 17:01:19 crc kubenswrapper[4739]: I1201 17:01:19.477801 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 17:01:19 crc kubenswrapper[4739]: E1201 17:01:19.478718 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 17:01:25 crc kubenswrapper[4739]: I1201 17:01:25.938061 4739 scope.go:117] "RemoveContainer" containerID="d96801347fdb8ec470810df69d89b6df74bb78c8a5017919f5e8a0b3e243de7f" Dec 01 17:01:25 crc kubenswrapper[4739]: I1201 17:01:25.981670 4739 scope.go:117] "RemoveContainer" containerID="b241eb5cc2912523531f85051bfb8a45822701881f4404121466cf7cd09697d1" Dec 01 17:01:26 crc kubenswrapper[4739]: I1201 17:01:26.024272 4739 scope.go:117] "RemoveContainer" containerID="1634a346c9fc7d2aece547cef373f74e710f0765ee07ddb4391c0b180fd6914f" Dec 01 17:01:33 crc kubenswrapper[4739]: I1201 17:01:33.478193 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 17:01:33 crc kubenswrapper[4739]: E1201 17:01:33.479142 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 17:01:47 crc kubenswrapper[4739]: I1201 17:01:47.477022 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 17:01:47 crc kubenswrapper[4739]: E1201 17:01:47.477898 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 17:01:58 crc kubenswrapper[4739]: I1201 17:01:58.484735 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 17:01:58 crc kubenswrapper[4739]: E1201 17:01:58.485758 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 17:02:13 crc kubenswrapper[4739]: I1201 17:02:13.478402 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 17:02:13 crc kubenswrapper[4739]: E1201 17:02:13.479269 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 17:02:24 crc kubenswrapper[4739]: I1201 17:02:24.477283 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 17:02:24 crc kubenswrapper[4739]: E1201 17:02:24.478114 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 17:02:35 crc kubenswrapper[4739]: I1201 17:02:35.477048 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 17:02:35 crc kubenswrapper[4739]: E1201 17:02:35.477824 4739 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rnmn7_openshift-machine-config-operator(dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e)\"" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" podUID="dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e" Dec 01 17:02:49 crc kubenswrapper[4739]: I1201 17:02:49.477202 4739 scope.go:117] "RemoveContainer" containerID="d5a7eb8d2fb3be289f734f7b890008bb5ab536d5ee9c496c079cf76dd0b56db6" Dec 01 17:02:50 crc kubenswrapper[4739]: I1201 17:02:50.154985 4739 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rnmn7" event={"ID":"dbc6c0ba-3636-4e9e-a4cd-b5fd0cdb597e","Type":"ContainerStarted","Data":"e109786130c1e7b04a8f58176319e9567209da5ca436634cd0f7b8ec664e7db7"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515113345155024450 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015113345156017366 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015113332335016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015113332336015455 5ustar corecore